Next Article in Journal
Picard-Jungck Operator for a Pair of Mappings and Simulation Type Functions
Previous Article in Journal
On a Variational Method for Stiff Differential Equations Arising from Chemistry Kinetics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Optimized Network Representation Learning Algorithm Using Multi-Relational Data

1
School of Computer, Qinghai Normal University, Xining 810800, China
2
Tibetan Information Processing and Machine Translation Key Laboratory of Qinghai Province, Xining 810008, China
3
Key Laboratory of Tibetan Information Processing, Ministry of Education, Xining 810008, China
4
School of Computer Science, Shaanxi Normal University, Xi’an 710119, China
*
Author to whom correspondence should be addressed.
Mathematics 2019, 7(5), 460; https://doi.org/10.3390/math7050460
Submission received: 26 April 2019 / Revised: 16 May 2019 / Accepted: 16 May 2019 / Published: 21 May 2019
(This article belongs to the Section Network Science)

Abstract

:
Representation learning aims to encode the relationships of research objects into low-dimensional, compressible, and distributed representation vectors. The purpose of network representation learning is to learn the structural relationships between network vertices. Knowledge representation learning is oriented to model the entities and relationships in knowledge bases. In this paper, we first introduce the idea of knowledge representation learning into network representation learning, namely, we propose a new approach to model the vertex triplet relationships based on DeepWalk without TransE. Consequently, we propose an optimized network representation learning algorithm using multi-relational data, MRNR, which introduces the multi-relational data between vertices into the procedures of network representation learning. Importantly, we adopted a kind of higher order transformation strategy to optimize the learnt network representation vectors. The purpose of MRNR is that multi-relational data (triplets) can effectively guide and constrain the procedures of network representation learning. The experimental results demonstrate that the proposed MRNR can learn the discriminative network representations, which show better performance on network classification, visualization, and case study tasks compared to the proposed baseline algorithms in this paper.

1. Introduction

Network representation learning (NRL) plays a pivotal role in many areas, which aims at learning low-dimensional, compressed, and distributed representation vectors for all kinds of networks. Network representation learning tasks can be intuitively regarded as network encoding tasks, where each node is given a unique vector in representation space, consequently, the neighboring vertices possess a closer space distance based on the distance evaluation function. Distributed representation learning is derived from a language-embedding application. The representative algorithm of language embedding is Word2Vec, proposed by Mikolov et al. [1,2]. Motivated by the Word2Vec algorithm, the network representation learning algorithm was consequently proposed by Perozzi [3], who adopted the random walk strategy to generate the vertex sequences, which are the same as “sentences” in language models. DeepWalk receives multiple vertex sequences v 0 , v 1 , v n as the input of the network representation learning model, where n denotes the length of the random walk sequence. In the middle layer of the learning model, the sequence v 0 , v 1 , v n is converted into the vertex pairs ( v 0 , v 2 ) , ( v 0 , v 1 ) , ( v 0 , v 1 ) , and ( v 0 , v 2 ) , where the window size needs to be set and the number of vertex pairs is controlled by the window size. The output of DeepWalk is a low-dimensional vector, r v k , where k is the size of the network representation vector. Recently, the DeepWalk algorithm has been successfully applied to many tasks [4].
There exist various relationships between network vertices. The structural relationships can be captured by DeepWalk and the community relationships can be captured by community detection algorithms or vertex labels. The vertex relationships in real-world networks also can be intuitively captured by the rich vertex contents (texts). For example, different titles of network vertices may contain the same words in citation networks, so the higher the overlap of words, the greater the probability of reference between them. Indeed, it is also well acknowledged that most existing NRL models ignore the semantic relationships between vertices. However, some NRL models try to integrate node contents into network representation learning. For example, text-associated deep walk (TADW) [5] and tri-party deep network representation (TriDNR) [6] provide a novel NRL framework, which integrates text features into representations by matrix factorization and joint learning. More specifically, TADW adopts inductive matrix completion (IMC) [7] to integrate text features into a network representation model; the essence of TADW is a kind of matrix factorization. Thus, it is not suitable for tasks of large-scale network representation learning. Two neural networks are applied by TriDNR to integrate text contents and label information from network vertices into a network representation learning model, which learns network representations in three parts (i.e., node structure, node content, and node label), consequently, TriDNR captures the inter–node, node–word, and label–word relationships. Because it considers the relationships between inter–node, node–word, and label–word, TriDNR has a higher computational complexity, and it would take a longer training time for representation learning, even in some smaller networks. Therefore, the objective of this paper is to research how to make full use of the network node text contents and how to improve the performance of network representation learning by node text contents. Finally, we hope that the proposed algorithm can obtain smaller computational complexity in tasks of large-scale network representation learning and can also gain a higher precision and desirable performance.
Multi-relational data modeling was first proposed by Bordes et al. [8], and it is a directed graph containing entities and relationships such as triplet form (head, label, tail). The triplet is denoted in the form ( h , l , t ) in this paper. An example is knowledge bases, such as Freebase [9] and DBpedia [10], where each entity is an abstract concept and the relationships are the facts involving two entities. TransE [8] can model multi-relational data using more generic methods that take all heterogeneous relationships into account at the same time. The TransE algorithm mainly learns knowledge embedding based on Freebase and DBpedia, etc.
The feature matrix of the network is factorized by TADW to gain the node representations, where the text feature matrix is as an auxiliary matrix. Note that the node contents are converted to a text feature matrix in TADW. This kind of TriDNR algorithm considers words in node contents as kind of special vertices and models node–word and word–label relationships, where all words in node contents are input into the neural network to train and learn the network representation vectors. Thus, we want to propose a novel network representation learning algorithm, which can model inter–vertex and vertex–word relationships with a lower calculation complexity. More importantly, the node representations can possess semantic information. To achieve the above goal, we first constructed the vertex relationships by node contents in this paper. We then converted these vertex relationships into vertex triplets in the form ( h , l , t ) , where l denotes the semantic relationship between vertex h and vertex t . Consequently, we propose an optimized network representation learning algorithm using vertex triplets, called MRNR, which focuses on modeling the network structure and multi-relational data at the same time. Motivated by the TransE model, we introduce the idea of the TransE algorithm to optimize the network representation learning process and adopt DeepWalk to model network structure features. Please note that we propose a new approach based on DeepWalk to model all knowledge triplets. Through this we learn the representations of triplets ( h , l , t ) using the improved algorithm without TransE. The TransE [8] and DeepWalk [3] are all proposed based on the Word2Vec [1,2]. So, we propose a unified joint learning framework and objective function without concatenating two kinds of vertex representations. In this framework, MRNR learns the network representations and model the multi-relational data using a single neural network (TriDNR uses two neural networks). Moreover, only some words in the node text contents are involved in the tasks of network representation learning. Therefore, MRNR is competent to the tasks of the large-scale network representation learning. Finally, we introduce a kind of optimization algorithm to improve the performance of network representation learning. However, MRNR is a novel node semantic relationship preserving network representation algorithm. By jointly optimizing the DeepWalk structure preserving objective and TransE [8] node semantic relation preserving objective, the proposed MRNR algorithm learns informative node representations.
To summarize, we make the following contributions:
(1)
We propose a novel vertex triplet-based NRL model, MRNR. It incorporates the multi-relational data between vertices into the network representation learning model. As a simple and straightforward model, it models the network representations between vertices and vertex triplets with a unified joint learning framework at the same time.
(2)
To the best of our knowledge, we are the first to introduce the idea of a vertex triplet into network representation learning, which means that the external multi-relational data are introduced into DeepWalk as a regularization term to guide and constrain the training procedure. Moreover, we introduce a kind of higher order transformation strategy to optimize the learnt network representation vectors.
(3)
We conduct experiments on the three real-world datasets with the task of node classification, the experimental results show that the MRNR is an effective and robust network representation approach. Importantly, MRNR can achieve excellent performance in large-scale network learning tasks, especially in terms of accuracy and time complexity.

2. Related Work

Representation learning [11] algorithms aim at avoiding huge feature engineering and improving the application areas of network features. Representation learning algorithms learn features via neural network or matrix factorization methods. DeepWalk [3] is a network representation learning algorithm based on Word2Vec [1,2]. DeepWalk [3] can encode complex network structures into the low-dimensional representation vector space, showing better network classification performance on various network tasks.
Consequently, some improved algorithms are proposed for network representation learning tasks. Several are based on DeepWalk [3], such as node2vec [12], NEU [13], WALKLETS [14] and HARP [15] etc., while node2vec [12] optimizes the strategy of random walk. For example, node2vec [12] adopts breadth first search and depth first search to sample the node sequences. The breadth first search provides a local microscopic view and depth first search provides a global macroscopic view. DeepWalk [3] can capture node relationships of lower order, and NEU [13] introduces a novel approach, which can obtain higher order network representations using simple representation transformation. Moreover, the NEU [13] algorithm implicitly approximates higher order proximities with theoretical approximation bound, WALKLETS [14] obtains higher order representations by the k-step random walk strategy, and HARP [15] learns higher order representations by graph compression strategy. The low order features are just to consider the local relations amongst vertices, and the high order features mainly take the global relations amongst vertices into account.
Another kind of improvement is based on neural matrix factorization of DeepWalk [3], such as Line [16], GraRep [17], TADW [5], MMDW [18]. Specifically, the Line [16] fully researches the 1st order proximity and the 2nd order proximity based on matrix factorization. DeepWalk [1] actually aims at factorizing the structure feature matrix of the networks. Consequently, GraRep [17] introduces a matrix factorization strategy to obtain higher order representations of the networks. GraRep [17] obtains higher order features of networks through multiplication of adjacency matrix. The uses inductive matrix completion (IMC) [7] is adopted by TADW [5] to integrate the node contexts into network representations. The max-margin method is used by MMDW [18] to optimize the learnt network representations. In fact, both TADW [5] and MMDW [18] adopt the matrix factorization approach to obtain the network representations.
The third kind of improvement is based on the joint learning model, such as DDRW [19], TriDNR [6], GCN [20], M-NMF [21], AANE [22], COSINE [23], and comE, [24] etc. Specifically, the DDRW is an improved algorithm based on DeepWalk [3], and DDRW [19] jointly improves the objective of classification and the objective of embedding entities. Two neural networks are adopted by TriDNR to learn the network representations from three parts (the node’s structure, node’s content, and node’s label), consequently, TriDNR [6] captures the inter–node, node–word, and label–word relationships. Compared with TADW [5], the relationships of the node–label could be learnt based on these two neural networks. Both the local graph structures and features of the nodes can be encoded by GCN [20] by learning the hidden layer representations. And M-NMF [21] is a modularized non-negative matrix factorization approach, which incorporates the community structures into network representations. Both COSINE [23] and comE [24] also incorporate the community information into network representations. In addition, AANE [22] is an accelerated attributed network embedding approach, which enables the joint learning framework to be conducted by decomposing the complicated learning and training into many sub-problems.
There exist some network representation approaches for special network structures, such as DynamicTriad [25], DepthLGP [26], and DHNE [27], etc. In addition, motivated by generative adversarial networks (GANs) [28], some network representation learning algorithms introduce GANs to optimize the learning procedure, such as ANE [29], GraphGAN [30], and NetGan [31].
To sum up, various solutions and novel algorithms are proposed for tasks of network representation learning. The focus of this paper is to research how to efficiently use network node contents to improve the performance of network representation learning. More importantly, we research how to use the network node contents to conduct learning and training tasks of network representation on large-scale networks, which can overcome higher calculation complexity like TADW [5] and TriDNR [6] algorithms. We mainly use TADW [5] and MRNR [6] proposed in this paper to make a comparative analysis in a fair way, as these algorithms only adopt node structures and node contents to learn network representation vectors.

3. Method

The MRMR algorithm proposed in this paper aims to embed the text features of network vertices into the network representation vectors. Some network representation learning methods based on matrix factorization take the text features of network vertices as the feature matrix to optimize the factorization procedures of the network structure feature matrix, for example, the TADW algorithm belongs to this kind of NRL algorithm. Other network representation learning methods based on neural networks simultaneously model the relationships between vertices and the relationships between vertices and text through multiple neural networks, for example, the TriDNR algorithm belongs to this kind of NRL algorithm. Different from the TADW and TriDNR algorithms, a novel text-embedding method was adopted in this paper, namely, the text between network vertices was transformed into the triplets between vertices. In order to embed vertex triplets into the network representation learning task, we introduced the idea of multi-relationship modeling, namely, we adopted the idea of knowledge-embedding learning using TransE to model the triplet relationships between network vertices. In order to constrain the modeling of network vertex relationships by modeling vertex triplet relationships, we propose the objective function of the MRNR algorithm based on the CBOW model optimized by the negative sampling and TransE algorithms. Therefore, we first introduce how to construct the triplets between vertices in this section, then introduce the modeling procedure of the TransE algorithm, and then we introduce the CBOW model based on negative sampling optimization, and finally we introduce how the MRNR algorithm proposed in this paper can simultaneously model the network vertex relationships and triplet relationships.
In Section 3.1, we introduce some definitions of the parameters used in this paper. In Section 3.2, we introduce how to construct the vertex triplets based on text contents of two vertices. In Section 3.3 and Section 3.4, the TransE algorithm and CBOW algorithm are introduced. After that, we propose the MRNR algorithm based on the improvements of these two algorithms in Section 3.5. Therefore, we introduce in detail how to embed vertex triplets into the network representation learning procedure based on the idea of the TransE algorithm in Section 3.5, and how to optimize the learnt network representation vectors obtained from the MRNR algorithm in Section 3.6.

3.1. Formalization

The parameters used in this paper are defined as follows. For a network G = ( V , E ) , we denote V as the vertex set, E as the edge set, and | V | as the size of vertex set, where each vertex v 0 , v 1 , v n V and each edge e 0 , e 1 , e m E . The purpose of the MRNR is to train the representation r v k for a given vertex v V , which is also the goal of network representation learning. Here, k is the size of network representation vector and it is expected to be much smaller than | V | . Namely, | V | usually denotes the number of vertices in the network, and we often set the size of the network representation vector as 100, 200, 300, etc., which is much smaller than the number of vertices. Note that r v k is not task-specific and it can be used for various tasks. Because of the output of the DeepWalk algorithm, so can the learnt network representation vectors r k be used for network node classification, network visualization, link prediction, and recommendation systems, etc. For the multi-relational data modeling, we set the ( h , l , t ) as vertex triplets, where h denotes the head entity, t denotes the tail entity, and l denotes the relations between head entity and tail entity. For the CBOW model in DeepWalk, we set the C o n t e x t ( v ) as the context vertices of the current vertex v and N E G ( v ) denotes the negative sampling procedure for the current vertex v . Here, the high-frequency variables are defined and explained, and the other variables adopted by this paper will be explained in detail at the corresponding sections.

3.2. Triplet Construction

The semantic relationships between vertices are extracted from the vertex texts. In this paper, we take two kinds of strategies to denote the relationships between vertices. The first kind of relationship is called as Unitary Relationship, where the relationship is constructed by one same word amongst word sequences from two vertex’s texts. The second kind of relationship is called as Binary Relationship, where the relationship is constructed by two same words amongst the word sequences from two vertex’s texts. We give a simple example in Table 1.

3.3. TransE Model

Inspired by Word2Vec [1,2], Bordes et al. [8] proposed the TransE algorithm, and the TransE algorithm models the relationships by training the triplets (head, label, tail) (denoted by ( h , l , t ) ) of multi-relational data as a form of translation tasks. It means that TransE needs to continuously adjust the representations of parameters h , l , and t to get a global equilibrium, h + l t . The objective function of TransE is as follows:
L = ( h , l , t ) S ( h , l , t ) S ( h , l , t ) [ γ + d ( h + l , t ) d ( h + l , t ) ] +
where
S ( h , l , t ) = { ( h , l , t ) | h E } { ( h , l , t ) | t E } .
Here, S is the set of triplets ( h , l , t ) and S is the set of negative samples. The energy of a triplet is equal to d ( h + l , t ) for the dissimilarity measure and d , γ > 0 is a margin hyperparameter. [ x ] + is the positive part of variable x .

3.4. CBOW Using Negative Sampling

DeepWalk can be implemented by CBOW and Skip-Gram introduced from Word2Vec, which all can be optimized by negative sampling (NEG) and Hierarchical Softmax [1,2]. In terms of training efficiency, CBOW and negative sampling outperforms Skip-Gram and Hierarchical Softmax. Thus, the CBOW using NEG is applied in our model.
For current vertex v , we define its context vertices as C o n t e x t ( v ) , the negative sampling procedure is defined as N E G ( v ) . Here, N E G ( v ) ϕ . For u D , D is the vertex set. We define the label as follows:
L v ( u ) = { 1 , u = v 0 , u v
where L v ( u ) is the sampling label of vertex u , for example, the label of positive sampling is 1, and the label of negative sampling is 0.
For one vertex and its contexts ( v , C o n t e x t ( v ) ) , our aim was to maximize the following probability in the entire representation space:
g ( v ) = ξ { v } N E G ( v ) p ( ξ | C o n t e x t ( v ) )
where
p ( ξ | C o n t e x t ( v ) ) = { σ ( x v T θ ξ ) L v ( ξ ) = 1 1 σ ( x v T θ ξ ) L v ( ξ ) = 0 ,
and σ ( x ) is the Sigmoid function, x v is the summing operation of each representation in C o n t e x t ( v ) , θ ξ is an assistant training vector of the current vertex ξ . Equation (5) can also be formalized as follows:
p ( ξ | C o n t e x t ( v ) ) = [ σ ( x v T θ ξ ) ] L v ( ξ ) · [ 1 σ ( x v T θ ξ ) ] 1 L v ( ξ )
so, g ( v ) can be calculated based on Equations (4) and (6) as follows:
g ( v ) = ξ { v } N E G ( v ) [ σ ( x v T θ ξ ) ] L v ( ξ ) [ 1 σ ( x v T θ ξ ) ] 1 L v ( ξ ) .
Therefore, for a given corpus C , the function:
G = v C g ( v )
is defined as the overall objective function, the objective function of CBOW using NEG is denoted based on Equations (6)–(8) as follows:
( v ) = log G = log v C g ( v ) = v C log g ( v ) = v C ξ { v } N E G ( v ) { L v ( ξ ) · log [ σ ( x v T θ ξ ) ] + ( 1 L v ( ξ ) ) · log [ 1 σ ( x v T θ ξ ) ] } .
We optimize Equation (9) by the stochastic gradient ascent algorithm. Consequently, the updating of θ u is computed as follows:
θ ξ : = θ ξ + μ · [ L v ( ξ ) σ ( x v T θ ξ ) ] x v
The updating approach of each context vertex v ( u ) is calculated as follows:
v ( u ) : = v ( u ) + μ ξ { v } N E G ( v ) [ L v ( ξ ) σ ( x v T θ ξ ) ] · θ ξ
where u C o n t e x t ( v ) .

3.5. MRNR Model

On citation networks, the relationships between vertices have reference relationships and referenced relationships. Generally, we just considered these relationships as a kind of edge existence. Some network representation learning algorithms based on shadow neural networks focus on learning the context relationships between the current vertex and context vertices. Consequently, the learnt representations tend to show the similarity if two vertices almost possess the same structure. Network vertices usually contain rich text contents on citation networks, such as paper title, abstract, and running title, etc. Empirically, the vertex without reference relationships maybe contain the same words in their titles, so we can construct unitary relationships and binary relationships by extracting one same word or two same words in titles. Therefore, we propose an optimized network representation learning algorithm using multi-relational data (MRNR) inspired by, literature [32,33], which jointly learns the network representations from the network representation learning model and knowledge representation learning model using TransE. Roughly speaking, MRNR integrates multi-relational data (triplets) into the procedure of learning the network representations. The detailed training diagram is shown in Figure 1.
As shown in Figure 1, the left training procedures are based on CBOW modeling and the right training procedures are based on multi-relation data modeling. For the current vertex 3, its context vertices are 1, 2, 4, and 5, in addition, there exists some unitary relationships and binary relationships between vertex 3 and 6. For all vertices, two kinds of learning models share common representation space, for the relationships between vertices, MRNR creates a new representation space, which can avoid conflicts between vertex representations and relationship representations.
Figure 1 shows the detailed model training procedures and the following contents would introduce the objective function of MRNR.
Equation (9) is the objective function of DeepWalk using CBOW and negative Sampling, it then can be simplified as follows:
( v ) = v C log g ( x )
The MRNR algorithm introduces the triplets ( h , l , v ) into DeepWalk, meanwhile, triplets ( h , l , v ) should satisfy h + l v which means that v should possess a nearest vector space distance with h + l , otherwise, h + l should possess a further vector space distance with v . We get inspiration from the TransE algorithm, namely, if ( h , l , v ) holds, then the representation of the vertex v should be closed to the representation of the vertex h plus the representation of the relationship l .
Motivated by the TransE model, we finally use the DeepWalk algorithm to model all the triplets. Consequently, we propose a new approach based on DeepWalk to learn the relationships of the triplet ( h , l , v ) . The objective function is as follows:
( v , l ) = v C l R v log g h + l ( v ) ,
where R v is the all triplets containing v . We can build the relationship between CBOW modeling and multi-relational modeling. In the MRNR model, we can regard the triplet ( h , l , v ) as a regularization term constraint to DeepWalk. To propose a joint learning framework, we define our objective function as follows:
= ( v ) + α ( v , l ) .
Here, we introduce the hyper-parameter α to balance the weights between structural features and triplet features.
For Equation (13), it can be replaced by Equation (9):
( v , l ) = v C l R v ξ { v } N E G ( v ) { L v ( ξ ) · log [ σ ( x h + l T θ ξ ) ] + ( 1 L v ( ξ ) ) · log [ 1 σ ( x h + l T θ ξ ) ] }
where x h + l = x h + x l . We use the stochastic gradient ascent algorithm to get the updating expressions of each parameter as follows:
θ ξ : = θ ξ + μ · [ L v ( ξ ) σ ( x h + l T θ ξ ) ] · x h + l
x h : = x h + μ ξ ( v } N E G ( v ) [ L v ( ξ ) σ ( x h + l T θ ξ ) ] · θ ξ
x h : = x h + μ ξ ( v } N E G ( v ) [ L v ( ξ ) σ ( x h + l T θ ξ ) ] · θ ξ
In Equation (14), we add the weight β for the triplet relation modeling. Therefore, we need to add a weight β to the Equations (16)–(18) when the variables are updated, namely, we replace the variable η with the variable η × β .

3.6. Higher Order Proximity Approximation of MRNR

In Line [16] algorithm, the neighboring vertices are captured to provide the first order similarity and the second order similarity between vertices, so it provides two kinds of models, namely, 1st Line and 2nd Line. In addition, GraRep [17] adopts the form of adjacency matrix multiplication to capture higher order features. Moreover, WALKLETS [14] captures higher order vertex relationships on random walk sequences, then it puts all vertex relationships of lower order and higher order into the DeepWalk model, generating effective and excellent network representations. However, these algorithms possess a higher computation complexity. Yang et al. [13] propose a unified framework of higher order proximity approximation, NEU, which can approximate a higher order proximity matrix for learning better network representations. Importantly, NEU [13] applies arbitrary network representations trained by lower order proximities as inputs of NEU to avoid the repeated computation.
The optimization expression of NEU is as follows:
R e = R e + λ 1 A · R e + λ 2 A · ( A · R e )
where R e is the network representation vectors of lower order proximities, A is the normalized adjacency matrix, and R e is the network representation vectors of higher order proximity. Therefore, NEU can implicitly approximate higher order proximities based on a theoretical approximation bound. Therefore, we use the NEU algorithm to enhance the performance of MRNR, which is a kind of optimization for the learnt representation vectors of MRNR.
The algorithm framework proposed in this paper is not a task-specified algorithm for NLP and NRL tasks. Consequently, MRNR uses DeepWalk to verify the feasibility of the proposed framework. In network representation learning algorithms, some network representation learning algorithms are based on the framework of neural network, and others are based on matrix factorization. And MRNR uses the basic DeepWalk algorithm to construct a joint learning framework. Afterwards, some improved algorithms have been proposed based on the DeepWalk algorithm, such as node2vec [12], MMDW [18], and SDNE [19]. Therefore, MRNR is a flexible framework which also can be applied to other NRL tasks.
In order to demonstrate the modeling procedure of the MRNR algorithm in more detail, we give the algorithm pseudocode, which does not add the procedure of higher order optimization, the pseudocodes of Algorithm 1 are shown as follows:
Algorithm 1: MRNR Algorithm
Input: graph: G(V, E), network representation vector size: k, random walk number: wn, random walk lengh: rl, triplet set: S, the weight of triplet: γ .
Output: vertex representation vectors of MRNR
// get all walk sequences of network vertices
1: for i form 1 to wn do
2:   O ← Shuffle(V)
3:   for v in O do
4:      C ← RandomWalkSequenceAppend(G, v , rl)
5:   end for
6: end for
  //initialize all variables
7: vertex_amount ← GetVocabAmount(C)
8: relation_amount ← GetRaltionAmount(S)
9: Vec Init_Representation_Vector(vertex_amount, k)
10:  θ Init_Representation_Vector(vertex_amount, k)
11:  P Init_Representation_Vector(relation_amount, k)
12: for v in C do
13:     e 0
14:     x v ξ C o n t e x t ( v ) Vec ( ξ )
15:    for ξ { v } N E G ( v ) do
16:     q σ ( x v T θ ξ )
17:        g μ · [ L v ( ξ ) q ]
18:        ee + g θ ξ
19:        θ ξ θ ξ + g x v
20:    end for
21:    for ξ   in   C o n t e x t ( v ) do
22:       Vec ( ξ ) Vec ( ξ ) + e
23:    end for
24:    for l in R v do
25:       e 0
26:       (h, l, v ) ← ObtainTriplets(R, l, v )
27:      for ξ = { v } N E G ( v ) do
28:         x h + l Vec ( h ) + P ( l )
29:        q σ ( x h + l T θ ξ )
30:        g μ · [ L v ( ξ ) q ]
31:        ee + g θ ξ
32:         θ ξ θ ξ + γ g x h + l
33:     end for
34:     Vec(v) ← Vec(v) + γ e
35:     P(l) ← P(l) + γ e
36:   end for
37: end for

4. Experiments and Evaluations

4.1. Dataset Setup

We conducted our experiments on the Citeseer (M10), Database System and Logic Programming (DBLP) (V4), and Simplified DBLP (SDBLP) datasets. The descriptions of the three datasets are given in Table 2:
The separate vertices are not connected with the other vertices. Therefore, the random walker cannot walk to the separate vertices for the DeepWalk, node2vec, and Line algorithms. For the GraRep algorithm, it was also impossible to learn the relationships between the separate vertices and other vertices, namely, the adjacency matrix was multiplied several times; however, the values in the multiplied matrix remained 0. Therefore, the NRL algorithm will get a poor performance for the sparse network containing separate vertices. Nevertheless, the NRL algorithm based on structures and vertex texts can gain satisfactory classification performance, such as the TADW algorithm. In order to fairly compare the performance of the above algorithms, we deleted all separate vertices on the Citeseer dataset as well as the DBLP dataset. The SDBLP dataset is a simplified dataset of DBLP. On the SDBLP, we only retained those vertices containing three edges at least.
The unitary and binary relationships between vertices were applied to the MRNR to improve the performance of DeepWalk. We used Equation (13) to model all the representations of triplets. The amount is shown in Table 3.

4.2. Baseline Algorithms

DeepWalk: DeepWalk can learn network representations for large-scale networks. It is a basic algorithm of NRL, which is usually applied as a baseline algorithm. It provides two kinds of training models, such as CBOW and Skip-Gram, and it also provides two kinds of optimization methods, such as Hierarchical Softmax and negative sampling. In our experiment, we used the CBOW and negative sampling for DeepWalk.
Line: Line is a network representation learning method proposed by Tang et al. [16]. And then Line studies the problem of encoding very large information networks into low-dimensional representation spaces. This method proposes the concept of first order proximity and second order proximity. Consequently, Line provides two kinds of models, 1st Line and 2nd Line. We use the 2nd Line model in this paper.
node2vec: Node2vec also adopts the BFS and DFS strategies to capture higher order neighboring vertices. Meanwhile, node2vec defines a second-order random walk to balance the DFS and BFS. The random walk path of node2vec is controlled by a parameter. Unlike the random walk of DeepWalk, the probability from the current vertex to next vertex is not exactly equal.
GraRep: GraRep learns the network representations by probability transition matrix, the probability transition matrix of the first order is A = D 1 S , where S is the adjacency matrix, and D is the diagonal matrix. The probability transition matrix of k-th order is A k . In this paper, we set the k value to three, which means that GraRep could capture the neighboring vertices of the first order, second order, and third order.
Text Feature (TF): We took the vertex’s text feature matrix r v | V | × 100 as the 100 dimensional representations of the network vertices.
DeepWalk + TF: We concatenated two vectors from DeepWalk and text feature to obtain the 100 dimensional representations of network vertices.
DeepWalk as Matrix Factorization (MFDW): DeepWalk has proved that it factorizes the matrix M = ( A + A 2 ) / 2 . Here, we used the singular value decomposition (SVD) [34] algorithm to factorize the matrix M. The network representation vectors are calculated by the expression W = U · S 0.5 .
TADW: The TADW incorporates text features of the networks into network representations with the IMC [7]. And it factorizes the matrix M = ( A + A 2 ) / 2 , where A consists of the reciprocal of the degree value of the vertex.
Simplified TADW (STADW): Because TADW adopts normalization to optimize text features two times. The first optimization is conducted before SVD factorization of text features and the second optimization is performed after SVD factorization. In STADW, we deleted the optimization procedures for the text feature, because the MRNR proposed in this paper simply adopted the unitary and binary vertex relationships, which are not optimized by the normalization methods.
MRNR@1: MRNR uses the unitary relationships to optimize the performance of NRL.
MRNR@2: MRNR uses the binary relationships to optimize the performance of NRL.
MRNR@1 + NEU: the proposed MRNR uses the unitary relationships to optimize the performance of NRL, meanwhile, the learnt representations are optimized by NEU.
MRNR@2 + NEU: the proposed MRNR uses the binary relationships to the performance of NRL, meanwhile, the learnt representations are optimized by NEU.

4.3. Classifiers and Experiment Setup

We conducted our experiments on the Citeseer, DBLP, and SDBLP datasets. We used the network classification task to verify the performance of MRNR. The classifier uses SVM implemented by Liblinear [35]. In addition, we randomly took a portion of labeled data as a training set and the rest was a testing set. We evaluated the accuracy of the classification based on the different proportions of training sets. This proportion varied from 10% to 90%. For the above baseline algorithms, we set the dimension size as 100, walk length as 40, walk number as 10, window size as 5, negative number as 5, and minimum node frequency as 5. The minimum vertex frequency means that the algorithm will discard some vertices that appear less than five times in the random walk sequences. Note that the weight of the triplet was set to 0.5 in Section 4.4. We repeated our experiment 10 times and reported the average accuracy in the following experiments.

4.4. Experimental Results and Analysis

The classification results are shown in Table 4 (Citeseer), Table 5 (DBLP), and Table 6 (SDBLP). The MRNR consistently and significantly outperformed the most baseline algorithms on different training proportions, showing its feasibility on network vertex classification tasks.
From Table 4, Table 5 and Table 6, we derived the following interesting observations:
(1) the introduced MFDW is a matrix factorization form of the DeepWalk algorithm. The performance of the MFDW algorithm was better than that of DeepWalk for the three datasets. The performance of TF was better than that of DeepWalk for the Citeseer and DBLP datasets. In addition, the performance of TF was better than that of Line, achieving an almost same performance with DeepWalk on the SDBLP dataset. The main reason why MFDW can achieve a better performance is that all vertices in MFDW participate and construct the structural feature matrix. Nevertheless, the DeepWalk algorithm avoids accurately calculating the structure feature matrix. Because TADW is an algorithm based on the MFDW, which incorporates the vertex texts into network representations, therefore, TADW can obtain a better performance and outperform the other baseline algorithms. In addition, TADW uses the inductive matrix completion algorithm rather than a simple SVD factorization. Therefore, the performance of inductive matrix completion can greatly prompt the performance of TADW to a certain extent. The performance of DeepWalk + TF shows that it is difficult to obtain performance improvement by concatenating two representations from DeepWalk and text features. In addition, the experimental results have proved that modeling the triplet relationships generally improves the performance of the original DeepWalk. Indeed, the proposed learning framework in this paper can be applied to other types of models.
(2) The performance of the MRNR@1 was better than that of TADW, DeepWalk, node2vec, MFDW, Line, and TF on the Citeseer, DBLP, and SDBLP. The performance of the MRNR@2 was slightly inferior than TADW on the Citeseer and DBLP. However, the performance of the MRNR@1 and MRNR@2 was superior to that of TADW on the SDBLP. The average degrees of Citeseer, DBLP, and SDBLP were 2.57, 11.936, and 25.339. The mean differences between MRNR@1 and TADW were 4.64%, 2.33%, and 1.6% on the Citeseer, DBLP, and SDBLP. It demonstrates that the performance differences were decreasing with the growth of the average degree. On the SBDLP dataset, although all baseline algorithms almost had the same performance, MRNR@1 and MRNR@2 achieved the best performance.
(3) The classification performance of the MRNR@1 was better than that of the MRNR@2, which demonstrated that the more triplets between vertices, the better the performance of the MRNR. However, the number of triplets had little impact on the performance of network representation learning on the dense SDBLP dataset. In addition, the learnt network representations were optimized by the NEU algorithm, therefore, the NRL performance of the MRNR@1 and MRNR@2 achieved significant improvement on the Citeseer and DBLP datasets. On the DBLP dataset, the optimization effect of NEU was not as good as that on the Citeseer and DBLP.
(4) the introduced TADW algorithm and the proposed MRNR algorithm also used text features to enhance the performance of network representation learning. More specifically, TADW used matrix factorization to incorporate the text features into network representations, but the MRNR adopted vertex triplets constructed by text features to constrain the modeling procedure of the MRNR. In the MRNR algorithm, we set the minimum vertex frequency to 5, which means these vertices with the degree value of less than 5 will be deleted in the procedure of random walk sampling. The influence of this behavior is that vertices of small occurrence frequency are not be input into the model to learn as the context vertices, which are only trained when they are the current vertices. In the TADW algorithm, all the vertices take part in the training procedure. The purpose of setting the minimum vertex frequency is to speed up the convergence of the model. Therefore, deleting low-frequency vertices in contexts can help improve the performance of network representation learning. However, STADW is a simplified algorithm based on TADW, which deletes all optimizations of text features, consequently, the performance of STADW is inferior to that of TADW on three datasets.

4.5. Parameter Sensitivity

The proposed MRNR algorithm converts the texts of the network vertices into the triplet form, and then MRNR concatenates the objective of the triplet constraint to the that of DeepWalk. In order to research the effect of the triplet on NRL, the weight parameter α was added to the constraint item of the triplet in Equation (14). By setting a weight value that varies from 0.1 to 0.9, we counted the accuracy of the classification tasks under different proportions of training datasets that ranged from 10% to 90% for each fixed weight value, and we then calculated the average accuracy in the last column. Because the classification performance of the triplets consisted of unitary relationships, it was better than that of the triplets composed of binary relationships; therefore, we only considered the triplets of binary relationships in this experiment. The detailed results are shown in Table 7, Table 8 and Table 9.
As shown in Table 7, Table 8 and Table 9, the classification accuracy slightly varied within 1%, with the increasing weight of the triplet on the DBLP and SDBLP datasets. However, the fluctuation of classification accuracy was relatively clear on the Citeseer dataset. Compared with the triplet weight of 0.1, the classification accuracy of the triplet weight of 0.9 approximately improved by 5.89%. In addition, the Citeseer dataset was a sparse network, and the DBLP and SDBLP datasets were dense networks. In conclusion, we found that the weight of the triplet had a great influence on the representation performance of sparse network representation learning, but the weight of the triplet had little effect on the representation performance of the dense network representation learning. The main reason was that the sufficient structural features enable all kinds of algorithms to almost possess the same performance on dense networks. On sparse networks, the opposite was true.

4.6. Visualizations

In the previous sections, we used classification tasks to verify the performance of the MRNR on the Citeseer, DBLP, and SDBLP. In this section, we explore the clustering ability of the learnt representations. We first randomly chose four network categories on the Citeseer and three categories on the DBLP and SDBLP. Each category contained 150 vertices on the Citeseer and 200 vertices on the DBLP and SDBLP. We used the t-SNE [36] algorithm to visualize the learnt representation vectors. The visualization results are shown in Figure 2.
As shown in Table 2, the amount of triplets constructed by the unitary relations was more than that of the triplets constructed by binary relations, meanwhile, the amount of unitary relations was more than that of the binary relations on the Citeseer, DBLP, and SDBLP. Consequently, the visualization results show that the clustering ability of MRNR@1 is superior to that of MRNR@2 on the Citeseer and DBLP, and they almost achieved the same clustering performance on the SDBLP. The visualization experiments show that the number of triplets would affect the visualization and clustering performance on the sparse networks, namely, enough triplets would contribute to the feature learning. Consequently, we can obtain excellent representation vectors if the massive triplets are introduced into network representation learning on the sparse networks. On the dense networks, the influence of the number of triplets on visualization is negligible. Moreover, the results of visualization are just consistent with the results of classification in Table 4, Table 5 and Table 6. Therefore, the feasibility of the proposed MRNR algorithm is confirmed from these two experiments.

4.7. Case Study

The proposed MRNR can learn the network representations using multi-relational data. In this section, we take a vertex’s title as an example to conduct an instance study on the Citeseer dataset. The target title is “An Evolutionary Algorithm That Constructs Recurrent Neural Networks”. Based on DeepWalk, TADW, MRNR@1, and MRNR@2, we then found the most relevant vertices using the cosine similarity method. We obtained the top three nearest neighboring vertices in Table 10. The triplet weight was set as 0.5 in this experiment.
As shown in Table 10, TADW uses a text feature matrix to improve the performance of network representation learning and MRNR applies the knowledge-based triplets to improve the performance of network representation learning, consequently, both TADW and MRNR show text generality and similarity based on the learnt representations. For example, the titles generated by TADW, MRNR@1, and MRNR@2 contain the same word “Neural Networks” or “Evolutionary”. This kind of representations trained by MRNR can be used for various tasks, such as recommendation systems, webpage ranking, etc. This example demonstrates that TADW and MRNR learn the network representations by balancing the weight between the structural feature similarity and text similarity.

5. Conclusions

In this paper, we introduced the idea of knowledge representation learning into the network representation learning, which is that we used knowledge triplets to constrain the training procedure of network representation learning. Consequently, we proposed a new network representation learning algorithm, MRNR, which uses the vertex’s texts to construct unitary and binary triplets between vertices. To testify the performance of MRNR, we conducted classification, visualization, and case study experiments on the Citeseer, DBLP, and SDBLP datasets. We found that the greater the number of triplets between vertices, the higher the accuracy of network classification, and the same results were obtained in visualization tasks. In this paper, we specially compare the classification performance of MRNR with the TADW algorithm. We found that the classification performance of TADW was inferior to that of the MRNR algorithm on the Citeseer, DBLP, and SDBLP. The MRNR and TADW also adopted text features to improve the performance of network representation learning. The difference was that TADW converts text features into a text matrix and MRNR converts the text features into knowledge-based triplets. Finally, according to the application requirements, we should choose the appropriate algorithm between MRNR and TADW. The MRNR algorithm proposed in this paper is more suitable for the task of large-scale network representation learning using text features due to its effectiveness. For small-scale networks, the matrix factorization-based TADW algorithm is also a suitable approach.

Author Contributions

Conceptualization, Z.Y., K.Z. and Y.Z.; Methodology, Z.Y., H.Z.; Software, K.Z., Z.W.; Validation, Z.Y. and Y.Z.; Formal Analysis, Y.Z.; Investigation, Z.Y.; Resources, Z.Y.; Data Curation, K.Z.; Writing—Original Draft Preparation, Z.Y.; Writing—Review and Editing, Y.Z. and Z.W.; Visualization, Z.Y.; Project Administration, Z.Y.; Funding Acquisition, Z.Y.

Funding

This research was supported by the Fundamental Research Funds for the Central Universities, China (Grant Number: 2017TS045).

Acknowledgments

The authors wish to express their gratitude to the anonymous reviewers and the associate editors for their rigorous comments during the review process. In addition, the authors also would like to thank the experimenters in our laboratory for their great contribution toward data collection; they are Ke Zhang, Zhaoyang Wang and Yu Zhu.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. In Proceedings of the ICLR, Scottsdale, AZ, USA, 2–4 May 2013. [Google Scholar]
  2. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the NIPS, Las Vegas, USA, 9–13 December 2013; pp. 3111–3119. [Google Scholar]
  3. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the KDD, New York, NY, USA, 24–37 August 2014. [Google Scholar]
  4. Li, G.H.; Luo, J.W.; Xiao, Q. Predicting microRNA-disease associations using network topological similarity based on deepWalk. IEEE Access 2017, 5, 24032–24039. [Google Scholar] [CrossRef]
  5. Yang, C.; Liu, Z.Y.; Zhao, D.L.; Sun, M.S. Network representation learning with rich text information. In Proceedings of the AAAI, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  6. Pan, S.R.; Wu, J.; Zhu, X.Q.; Wang, Y. Tri-party deep network representation. In Proceedings of the IJCAI, New York, NY, USA, 9–15 July 2016. [Google Scholar]
  7. Natarajan, N.; Dhillon, I.S. Inductive matrix completion for predicting gene-disease associations. Bioinformatics 2014, 30, 60–68. [Google Scholar] [CrossRef] [PubMed]
  8. Bordes, A.; Usunier, N.; Garcia-Duran, A. Translating embeddings for modeling multi-relational data. In Proceedings of the NIPS, Harrah, NV, USA, 28 March–2 September 2013. [Google Scholar]
  9. Bollacker, K.; Evans, C.; Paritosh, P.; Taylor, J. Freebase: A collaboratively created graph database for structuring human knowledge. In Proceedings of the SIGMOD, Vancouver, BC, Canada, 9–12 June 2008. [Google Scholar]
  10. Lehmann, J. DBpedia: A large-scale, multilingual knowledge base extracted from wikipedia. Semant. Web 2015, 6, 167–195. [Google Scholar]
  11. Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pami 2013, 35, 1798–1828. [Google Scholar] [CrossRef] [PubMed]
  12. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the SIGKDD, San Francisco, CA, USA, 26 June–1 July 2016. [Google Scholar]
  13. Yang, C.; Sun, M.S.; Liu, Z.Y.; Tu, C.C. Fast network embedding enhancement via high order proximity approximation. In Proceedings of the IJCAI, Melbourne, Australia, 19–25 August 2017. [Google Scholar]
  14. Perozzi, B.; Kulkarni, V.; Skiena, H.C.; Skiena, S. Don’t walk, skip! online learning of multi-scale network embeddings. arXiv 2017, arXiv:1605.02115. [Google Scholar]
  15. Chen, H.C.; Perozzi, B.; Hu, Y.F.; Skiena, S. HARP: Hierarchical representation learning for networks. In Proceedings of the AAAI, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  16. Tang, J.; Qu, M.; Wang, M.Z.; Zhang, M.; Yan, J.; Mei, Q.Z. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015. [Google Scholar]
  17. Cao, S.S.; Lu, W.; Xu, Q.K. GraRep. Learning graph representations with global structural information. In Proceedings of the CIKM, Melbourne, VIC, Australia, 19–23 October 2015. [Google Scholar]
  18. Tu, C.C.; Zhang, W.C.; Liu, Z.Y.; Sun, M.S. Max-Margin DeepWalk: Discriminative Learning of Network Representation. In Proceedings of the IJCAI, New York, NY, USA, 9–15 July 2016. [Google Scholar]
  19. Li, J.Z.; Zhu, J.; Zhang, B. Discriminative deep random walk for network classification. In Proceedings of the ACL, Berlin, Germany, 7–12 August 2016. [Google Scholar]
  20. Kipf, T.N.; Welling, M. Semisupervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  21. Wang, X.; Cui, P.; Wang, J. Community Preserving Network Embedding. In Proceedings of the AAAI, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  22. Huang, X.; Li, J.D.; Hu, X. Accelerated attributed network embedding. In Proceedings of the SDM, Houston, TX, USA, 27–29 April 2017. [Google Scholar]
  23. Zhang, Y.; Lyu, T.S.; Zhang, Y. COSINE: Community-preserving social network embedding from Information diffusion cascades. In Proceedings of the AAAI, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  24. Cavallari, S.; Zheng, V.W.; Cai, H.Y.; Chang, K.C.C.; Cambria, E. Learning community embedding with community detection and node embedding on Graphs. In Proceedings of the CIKM, Singapore, 6–10 November 2017. [Google Scholar]
  25. Zhou, L.K.; Yang, Y.; Ren, X.; Wu, F.; Zhuang, Y.T. Dynamic network embedding by modeling triadic closure process. In Proceedings of the AAAI, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  26. Ma, J.X.; Cui, P.; Zhu, W.W. Depth LGP: Learning embeddings of out-of-sample nodes in dynamic networks. In Proceedings of the AAAI, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  27. Tu, K.; Cui, P.; Wang, X.; Wang, F.; Zhu, W.W. Structural deep embedding for hyper-networks. In Proceedings of the AAAI, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  28. Goodfellow, I.L.; Pouget-Abadie, J.K.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. In Proceedings of the NIPS, Palais des Congrès de Montréal, Montréal, QC, Canada, 8–13 December 2014. [Google Scholar]
  29. Dai, Q.Y.; Li, Q.; Tang, J.; Wang, D. Adversarial network embedding. In Proceedings of the AAAI, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  30. Wang, H.W.; Wang, J.; Wang, J.L.; Zhao, M.; Zhang, W.N.; Zhang, F.Z.; Xing, X.; Guo, M.Y. GraphGAN: Graph Representation Learning with Generative Adversarial Nets. In Proceedings of the AAAI, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  31. Bojchevski, A.; Shchur, O.; Zügner, D.; Günnemann, S. NetGAN: Generating graphs via random walks. In Proceedings of the ICML, Long Beach, CA, USA, 10–15 June 2018. [Google Scholar]
  32. Liu, Q.; Jiang, H.; Wei, S.; Ling, Z.H.; Hu, Y. Learning Semantic Word Embeddings based on Ordinal Knowledge Constraints. In Proceedings of the ACL, Beijing, China, 26–31 July 2015. [Google Scholar]
  33. Nguyen, K.A.; Walde, S.S.I.; Vu, N.T. Integrating Distributional Lexical Contrast into Word Embeddings for Antonym-Synonym Distinction. In Proceedings of the ACL, Berlin, Germany, 8–12 August 2016. [Google Scholar]
  34. Hofmann, T. Probabilistic latent semantic indexing. ACM SIGIR 1999, 51, 50–57. [Google Scholar]
  35. Fan, R.E.; Chang, K.W.; Hsieh, C.J. Liblinear: A library for large linear classification. J. Mach. Learn. Res. 2008, 9, 1871–1874. [Google Scholar]
  36. Maaten, L.V.D.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Learning framework of the proposed algorithm in this paper.
Figure 1. Learning framework of the proposed algorithm in this paper.
Mathematics 07 00460 g001
Figure 2. Visualization experiments. (ac) The MRNR@1′s visualization results on the Citeseer, DBLP, and SDBLP. (df) The MRNR@2′s visualization results on the Citeseer, DBLP, and SDBLP.
Figure 2. Visualization experiments. (ac) The MRNR@1′s visualization results on the Citeseer, DBLP, and SDBLP. (df) The MRNR@2′s visualization results on the Citeseer, DBLP, and SDBLP.
Mathematics 07 00460 g002
Table 1. Example of a unitary and binary relationship.
Table 1. Example of a unitary and binary relationship.
Node TextNode1: Online Learning of Multi-Scale Network Embedding
Node2: Large-Scale Information Network Embedding
Total
Unitary Relationship(1, Network, 2), (1, Embedding, 2)2
Binary Relationship(1, Network_Embedding, 2)1
Table 2. Dataset descriptions.
Table 2. Dataset descriptions.
DatasetsOriginal NodeOriginal EdgeSingle NodeUsed NodeUsed EdgeAverage Clustering CoefficientAverage Degree
Citeseer10,31059235700461059230.2642.57
DBLP60,744105,78143,01917,725105,7810.18711.936
SDBLP60,744105,7810311939,5160.25925.339
Table 3. Triplet and relationship count on the Citeseer, DBLP, and SDBLP.
Table 3. Triplet and relationship count on the Citeseer, DBLP, and SDBLP.
DatesetCiteseerDBLPSDBLP
TripletRelationshipTripletRelationshipTripletRelationship
Unitary Relationship1,263,566652415,848,32811,858510,7934451
Binary Relationship54,1333470115,540935657991264
Table 4. Accuracy (%) of vertex classification on the Citeseer dataset.
Table 4. Accuracy (%) of vertex classification on the Citeseer dataset.
Labeled Nodes (%)10%20%30%40%50%60%70%80%90%
DeepWalk55.8959.3060.8961.4862.1962.3062.6262.3363.95
Line42.6447.0648.0449.5750.4351.0251.1853.0753.63
node2vec62.4765.6166.0067.0767.4067.1567.4668.0768.56
GraRep39.3853.0957.8559.7559.9761.0561.5762.0960.89
Text Feature (TF)57.6961.3062.7663.0563.4863.3062.8762.1963.95
DeepWalk + TF58.3161.1562.7363.3764.1863.9665.5065.4965.30
MFDW57.6260.7962.3363.0562.9663.0063.0063.4864.30
TADW68.4772.2973.5074.2574.2474.7375.3674.9575.47
STADW46.7161.9965.9971.6073.1173.8273.3374.2774.36
MRNR@176.3177.8578.3378.4379.3179.3979.3779.6479.72
MRNR@266.7268.9370.1370.7170.8370.9471.6171.9171.50
MRNR@1 + NEU77.0578.5479.3979.4279.1479.5979.6980.3179.86
MRNR@2 + NEU67.4169.0270.1471.0671.7771.8872.1873.5371.76
Table 5. Accuracy (%) of vertex classification on the DBLP dataset.
Table 5. Accuracy (%) of vertex classification on the DBLP dataset.
Labeled Nodes (%)10%20%30%40%50%60%70%80%90%
DeepWalk62.2664.3465.4265.9866.2466.1866.6067.0366.77
Line64.4966.5367.4967.8767.9868.3069.0368.8968.86
node2vec73.3973.9875.2575.6175.7075.8575.7975.7376.36
GraRep58.9065.9067.2667.9268.7768.8869.2669.5669.79
Text Feature (TF)66.1769.4670.4971.1571.2971.4471.5471.5771.83
DeepWalk + TF62.6165.1565.9966.2266.3766.6067.0366.9167.61
MFDW65.0274.6874.8875.0275.0575.1375.2274.5775.51
TADW80.2281.4182.0582.3682.7782.8882.7183.1082.97
STADW75.3980.5181.0981.0081.6481.7081.7181.6981.34
MRNR@182.6183.9184.3984.7284.6584.9285.0185.4685.87
MRNR@277.3379.7680.6380.9881.1181.6881.6981.8181.58
MRNR@1 + NEU83.8685.7886.4486.2886.5686.5286.8186.9586.79
MRNR@2 + NEU80.3482.2982.7883.1383.5483.7683.6884.0184.88
Table 6. Accuracy (%) of vertex classification on the SDBLP dataset.
Table 6. Accuracy (%) of vertex classification on the SDBLP dataset.
Labeled Nodes (%)10%20%30%40%50%60%70%80%90%
DeepWalk79.7680.6581.8881.4982.5682.3582.7382.7183.37
Line73.7977.0178.1178.2879.3178.9779.6378.8278.77
node2vec82.1782.8783.0984.5184.4584.0184.8484.7384.98
GraRep80.9982.5284.1484.7884.9784.1785.3685.2784.95
Text Feature (TF)65.0371.2372.6473.8674.5475.0775.1476.0075.33
DeepWalk + TF79.3980.9581.6081.4482.5582.2282.1082.5883.15
MFDW79.7983.0884.3884.1284.5384.2984.7084.5584.53
TADW81.3883.1383.5484.8484.0984.3485.0085.6985.78
STADW80.6381.9282.5483.6084.0284.9285.0384.6585.27
MRNR@183.9284.7985.1585.3085.4385.8985.3286.5185.21
MRNR@283.8084.6185.2985.4185.3285.3486.0385.1585.46
MRNR@1 + NEU83.8484.8885.1885.5885.7586.3585.7385.6885.85
MRNR@2 + NEU83.4484.7484.8785.0185.2985.3685.6185.7185.11
Table 7. Influence of triplet weight on the Citeseer dataset.
Table 7. Influence of triplet weight on the Citeseer dataset.
α The Proportions of Training DatasetsAverage
10%20%30%40%50%60%70%80%90%
0.160.6763.6664.9765.2366.1266.6866.9066.8967.0965.36
0.264.1666.5168.1568.6368.8868.8768.6769.1270.3068.14
0.364.7069.0369.5669.7370.1071.2670.6070.7271.4369.68
0.466.5068.7269.7070.3770.8770.9970.9171.6371.4370.12
0.566.7268.9370.1370.7170.8370.9471.6171.9171.5070.36
0.667.2269.2170.3170.5971.2871.2571.9771.5372.0770.60
0.766.8669.4170.1670.8371.4672.1572.0972.3173.2270.94
0.866.7269.8070.5371.2771.7272.3272.7073.1973.3071.28
0.967.2869.3170.8171.7172.0772.1572.9772.2572.7271.25
Table 8. Influence of triplet weight on the DBLP dataset.
Table 8. Influence of triplet weight on the DBLP dataset.
α The Proportions of Training DatasetAverage
10%20%30%40%50%60%70%80%90%
0.177.9879.5880.3180.6280.9481.0881.4481.2881.1680.49
0.277.6279.4580.4480.7481.0581.0081.0780.9581.3980.41
0.377.2379.8080.4780.9080.8881.2581.1681.4782.2580.60
0.477.5280.0480.4780.8781.2981.7081.4581.6181.9080.76
0.577.3379.7680.6380.9881.1181.6881.6981.8181.5880.73
0.677.4979.7580.6180.8981.1881.4981.8781.8782.0880.80
0.777.5779.9480.7581.0681.3681.7081.6881.9481.8180.87
0.877.5579.9180.6180.9681.1581.4581.6581.9681.9680.80
0.978.0680.0480.7880.9581.2981.5181.7781.9982.3480.97
Table 9. Influence of triplet weight on the SDBLP dataset.
Table 9. Influence of triplet weight on the SDBLP dataset.
α The Proportions of Training DatasetAverage
10%20%30%40%50%60%70%80%90%
0.183.7484.6285.5285.3785.4485.3185.7585.6085.9285.25
0.283.9784.9784.9085.0284.9685.7785.4885.8685.7685.19
0.383.8285.1684.9685.2985.5485.4385.6785.3386.8585.34
0.483.8484.4185.3384.9385.2185.4786.1985.0786.2785.19
0.583.8084.6185.2985.4185.3285.3486.0385.1585.4685.15
0.683.9484.8785.1085.3985.4585.6586.2885.6285.5085.31
0.783.9484.7585.0385.0384.9385.7085.5485.6785.0285.07
0.883.7884.6385.2985.3785.5285.5585.8585.6586.5385.35
0.983.9384.7185.0385.0385.5085.3185.4886.2185.5685.20
Table 10. Case study on the Citeseer dataset.
Table 10. Case study on the Citeseer dataset.
AlgorithmTitleSimilarity
DeepWalkThe Evolution of Communication in Adaptive Agents0.8274
Challenges and Opportunities of Evolutionary Robotics0.8031
Applying Evolutionary Computation to Designing Neural Networks: A Study of the State of the Art 0.7998
TADWPareto Evolutionary Neural Networks0.7705
Compensating for Neural Transmission Delay using Extrapolatory Neural Activation in Evolutionary Neural Networks0.7491
An Adaptive Merging and Growing Algorithm for Designing Artificial Neural Networks 0.7449
MRNR@1Applying Evolutionary Computation to Designing Neural Networks A Study of The State of The Art0.9508
Making Use of Population Information in Evolutionary Artificial Neural Networks0.9447
Competitive Coevolution Through Evolutionary Complexification 0.9415
On the Combination of Local and Evolutionary Search for Training Recurrent Neural Networks0.9736
MRNR@2Applying Evolutionary Computation to Designing Neural Networks a Study of The State of The Art 0.9648
Evolutionary Reinforcement Learning of Artificial Neural Networks 0.9647

Share and Cite

MDPI and ACS Style

Ye, Z.; Zhao, H.; Zhang, K.; Zhu, Y.; Wang, Z. An Optimized Network Representation Learning Algorithm Using Multi-Relational Data. Mathematics 2019, 7, 460. https://doi.org/10.3390/math7050460

AMA Style

Ye Z, Zhao H, Zhang K, Zhu Y, Wang Z. An Optimized Network Representation Learning Algorithm Using Multi-Relational Data. Mathematics. 2019; 7(5):460. https://doi.org/10.3390/math7050460

Chicago/Turabian Style

Ye, Zhonglin, Haixing Zhao, Ke Zhang, Yu Zhu, and Zhaoyang Wang. 2019. "An Optimized Network Representation Learning Algorithm Using Multi-Relational Data" Mathematics 7, no. 5: 460. https://doi.org/10.3390/math7050460

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop