Next Article in Journal
On the Downlink Capacity of Cell-Free Massive MIMO with Constrained Fronthaul Capacity
Next Article in Special Issue
Mutual Information Based Learning Rate Decay for Stochastic Gradient Descent Training of Deep Neural Networks
Previous Article in Journal
Analysis of Multifractal and Organization/Order Structure in Suomi-NPP VIIRS Normalized Difference Vegetation Index Series of Wildfire Affected and Unaffected Sites by Using the Multifractal Detrended Fluctuation Analysis and the Fisher–Shannon Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Graph-Based Author Name Disambiguation Method and Analysis via Information Theory

1
School of Information Science and Technology, ShanghaiTech University, Shanghai 201210, China
2
Shanghai Institute of Microsystem and Information Technology, Chinese Academy of Sciences, Shanghai 200050, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
4
University of Science and Technology of China, Heifei 230026, China
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(4), 416; https://doi.org/10.3390/e22040416
Submission received: 6 March 2020 / Revised: 31 March 2020 / Accepted: 4 April 2020 / Published: 7 April 2020
(This article belongs to the Special Issue Deep Artificial Neural Networks Meet Information Theory)

Abstract

:
Name ambiguity, due to the fact that many people share an identical name, often deteriorates the performance of information integration, document retrieval and web search. In academic data analysis, author name ambiguity usually decreases the analysis performance. To solve this problem, an author name disambiguation task is designed to divide documents related to an author name reference into several parts and each part is associated with a real-life person. Existing methods usually use either attributes of documents or relationships between documents and co-authors. However, methods of feature extraction using attributes cause inflexibility of models while solutions based on relationship graph network ignore the information contained in the features. In this paper, we propose a novel name disambiguation model based on representation learning which incorporates attributes and relationships. Experiments on a public real dataset demonstrate the effectiveness of our model and experimental results demonstrate that our solution is superior to several state-of-the-art graph-based methods. We also increase the interpretability of our method through information theory and show that the analysis could be helpful for model selection and training progress.

Graphical Abstract

1. Introduction

Entity Linking tasks recognize or disambiguate named entities to an entity in a knowledge base. It is a significant problem in natural language processing and has been extensively studied. One important task of entity linking is author name disambiguation. The author name disambiguation task aims to partition publications written by different people who share the same name such that each partition only contains documents associated with one real-life person. In the field of bibliographic data analysis and document retrieval, author name disambiguation is crucial. For instance, when someone is looking for the publications of a scholar name “Charles” in a database, the query may return many papers from different “Charles”, which could cause ambiguity and deteriorate the performance of this search. The efficiency of the search could be significantly enhanced if the search results were distinguished by name references. Besides, if an organization want to calculate the impact of many authors, they need to know their publications exactly.
There are two types of questions that could be answered by solving author name disambiguation task. One is to find out who writes the specific paper and another is to decide which articles are written by this author. Author name disambiguation becomes more challenged when dealing with scholars not from western countries, as people with different names share the same spelling. For example, there are more than 300 thousand people named “Wei Zhang” in China.
Despite the fact that the problem has been studied for decades, it remains largely unsolved. Some researchers consider author name disambiguation as a classification task and manage to predict the correct label for each paper [1] or predict whether two documents are related to the same author [2]. A large number of labels are needed in classification tasks. Other works focus on clustering methods, which are usually unsupervised but lead to relatively poor performance. Graph networks have attracted more and more attention in recent years. As co-authors and their publications could naturally construct author-author network and document-document network, the author name disambiguation task could be solved by incorporating graph neural networks.
Handling name ambiguity problems presents several challenges:
  • Should we use labeled data? The disambiguation performance of supervised methods is generally better than unsupervised methods because of utilizing labeled data. But the size of datasets is usually large and manually labeling all tags will consume a lot of manpower and time. Solving this problem by semi-supervised methods has become a more practical solution to this problem.
  • How do we make the best of information and present a comprehensive model that could fit most datasets? Most existing methods are usually based on either feature extraction or on paper relationship and author relationship. Feature extraction methods adopt a large number of attributes and formulate many rules to measure the similarity between documents. When some attributes are missing, the rules will be invalidated. Research based on relationship graphs ignores some basic attributes of documents and reduces the performance of name disambiguation. It is necessary to design a basic method to deal with all kinds of datasets with different attributes.
  • Does the proposed method work well on larger datasets? Many state-of-art methods dealing with author name disambiguation are evaluated on relatively small document sets, associated with no more than twenty author name references. When the dataset becomes larger, the performance may drop a lot.
  • Why does the proposed method work well? Graph-based name disambiguation approaches can be interpreted as introducing relationship. We wish to know what role the training process plays and why some approaches lacking interpretability perform well.
In this paper, we propose a name disambiguation method that could generate document embedding vectors without labeled data and utilizes the true number of clusters for clustering.
Our method could make the best of the attributes of documents and the relationship of papers as well as the relationship of co-authors. Incorporating a word2vec model, we use basic attributes for feature extraction: title, author, organization and construct document representation vectors. The word2vec model could work on the dataset with some missing data and is flexible to involve other features. Two models are used for representation learning afterward. A Graph Auto-Encoder (GAE) is applied to construct a paper-paper network based on feature similarity and a Graph Embedding model is used to utilize the graph topology information of co-author relationship. Our method dealing with basic attributes could be easily applied to other author datasets.
A larger dataset sampled from AMiner [3] is used for evaluating the disambiguation results of our model. Experimental results indicate that our solution achieves significantly better performance than several state-of-the-art graph-based methods including Zhang and Yao [4], Zhang et al. [5] and GHOST [6]. Our method finds the document embedding vectors based on an unsupervised learning method and is even superior to the semi-supervised method [4].
Besides, we try to increase the interpretability of our method through information theory. Through the analysis of changes in mutual information, we could learn more about our model and the training progress.
In addition, visualization through dimensionality reduction is used for comparing our approach and other’s work and assessing the role of each component.
The remainder of this paper is organized as follows: In Section 2, we introduce recent related works on name disambiguation, including the supervised, unsupervised, semi-supervised and graph-based methods, and works on the information plane. In Section 3, we formalize the problem with preliminaries. The framework is presented in Section 4 and the calculation of mutual information is posed in Section 5. We evaluate our method in Section 6. In Section 7, we conclude the article.

2. Related Work

2.1. Name Disambiguation

A variety of methods have been proposed to solve the issue of author name ambiguity. Some researchers treat it as a classification task, predicting the correct labels of each paper or determining whether two articles are written by the same author. Classification tasks require lots of labels, so such tasks are usually supervised.
Wang et al. [1] proposed a boosted tree classification method for determining true papers and false papers of a set of homonymous authors in academic research processes. The input of the models is the attributes of publications, including title, author, affiliation, abstract, etc. A deep neural network model to learn attributes automatically is used in [2]. Some researchers have noticed the potentials of graph-based feature and utilized a Bi-directional Long Short-Term Memory to encode graph attributes for classification [7]. Many traditional methods measure similarity metrics by cosine similarity of term frequency-inverse document frequency (TD-IDF). Others train a model to learn a vector representation by triplets samples without considering any structure information. Kim et al. [8] proposed a hybrid model that makes use of both and train an SVM, a RF, a GBT and a DNN to determine whether a pair of publications is related to the same author or not.
Some works make use of external data collected from the web. A disambiguation Support Vector Machines (SVM) kernel is trained to leverage the rich structure and extensive coverage of the information encoded in Wikipedia in [9]. Han et al. [10] offered a generative Naive Bayes probability model and an SVM model. They illustrate these two approaches on data collected from the web and DBLP citation database (https://dblp.uni-trier.de).
In real digital libraries, the cost of human annotation for each name is unaffordable and supervised classification solutions are inefficient and inappropriate. Therefore, some unsupervised clustering models are applied to this task. For unsupervised name disambiguation, publications are divided into several clusters in order that each one only contains records associated with one person.
Some papers studied the efficiency of different clustering algorithms, such as Adaptive Threshold Clustering (ATC) [11] and Bayesian Information Criterion-mean (BIC-mean) [12]. Cen [13] modeled a new pairwise similarity measure by optimizing the linear regression model and then adaptively cluster documents by optimizing a regression function. A unified probabilistic framework and a hierarchical agglomerative clustering method based on Dempster–Shafer theory (DST) [14] propose new similarity metrics in name disambiguation task. They embed each document in a low dimensional vector space for clustering. Some probabilistic relational [15,16,17] models are also unsupervised.
The main drawback of supervised methods is that a large amount of labeled data requiring expensive human annotation is needed while unsupervised clustering methods usually perform worse than supervised solutions. The following methods are based on semi-supervised algorithms.
Levin et al. [18] proposed a two-stage method that combines the classification and clustering method. In stage one, they find positive document pairs that are probably written by the same researcher through high-precision rules, while articles that are not linked by the rules are negative ones. In stage two, the positive and negative document pairs are taken as training data for a supervised classifier that predicts whether two articles are related to the same author or not. Predictions of the classifier are then used as a similarity metric for clustering. Louppe et al. [19] posed phonetic-based blocking strategies for pre-clustering, a linkage function to learn an accurate classifier for identifying authors and semi-supervised agglomerative clustering of signatures. Blocking strategies are applied for placing documents that are likely to be written by the same author in the same area.
Some graph-based approaches are also used in this task due to the growing research interests in graph neural networks. Unsupervised learning approach using K-way spectral clustering [20] could divide a graph into several parts for clustering. Ref. [21] constructs a co-authors graph and applies graph structural clustering method ’gSkeletonClu’ to detect outliers and identify overlapping communities. GHOST [6] utilizes a coauthorship graph to compute the similarity between node pairs, utilizing the attribute of coauthor only could achieve the same performance as previous complicated approach. A combined graph encompassing the author-author graph and document-document graph is put forward by Pooja [22]. Each connected component of the combined graph represents a distinct cluster. In [23], researchers designed a deepwalk algorithm on the author-author graph and document-document graph and embedded graph nodes such that the document representations contain the information of the graph topologies. A pairwise factor graph (ADANA) [24] and title-coauthor graph (GFAD) [25] are also proposed.
In previous work [4], global metric learning and local linkage graph are combined. This method learns low-dimensional documents embedding vectors from their attribute characteristics. In global metric learning, the authors sampled real positive and negative pairs for triplet learning and utilized Graph Auto-Encoder in local graph learning. However, they did not consider the relationship between authors and used a lot of labeled samples. Zhang et al. [5] pre-processed the input data as three graphs: person-person graph, document-document graph, and person-document graph. They projected the data into low dimensional space. However, features other than co-author relationship are not taken into account and a lot of information is lost. Our proposed method takes advantages of the above two methods by leveraging both attributes and relationships.
Apart from author name disambiguation, some researchers are devoted to person name disambiguation in web search and the solutions are very similar to author name disambiguation. In [26], web pages are clustered after feature selection and measuring similarity as well.

2.2. Information Plane

The amount of information between a hidden layer and input/output can then be measured over training progress, yielding a picture of the optimization process in the information plane [27]. Works via information plane gives us a better understanding of neural networks.
Tishby [28] visualized the mutual information of hidden layers regarding input and output of a fully connected neural network via information plane and provided a better understanding of Deep Neural Networks (DNN) and the optimization process. They suggested that the goal of the network is to optimize the Information Bottleneck tradeoff between compression and prediction. The training progress could be divided into two phases by visualization. In phase one, hidden layers keep extracting information from input variables ( I ( X ; T ) increases during phase one while T represents random variables of hidden layers). In phase two, the layers remove irrelevant information ( I ( X ; T ) decreases).
Cheng et al. [29] verified the validation of information plane on classic Convolutional Neural Network including AlexNet and VGG. They learn the capability of these CNN structures based on the information plane.
On the one hand, we wish to know whether our graph networks still follows the process of extracting information and compression in the IB principle. On the other hand, the variation of mutual information I ( X ; Y ) could reflect information flow from a variable X to a variable Y. We calculate the mutual information of latent vectors regarding input and output to find out whether IB principle holds and increase the interpretability of our framework.

3. Preliminaries

3.1. Problem Formulation

In this section, we formalize the name disambiguation problem. Firstly, we introduce the notation used in this paper.
Let D be the set of all documents in the database, and | D | is the number of the documents. For an author name i, D i = { d 1 i , d 2 i , , d N i } is a set of | N | documents associated with the name. Each document has a set of attributes, like coauthor, title, venue, affiliation, keywords, etc. We let d n be the n-th document, d n could be represented as { f 1 , f 2 , , f M } , in which f is a stemmed word of each attribute and M stands for the number of stemmed words of these attributes. Let A be the collaborator set, then it can be expressed as A = { a 1 , a 2 , , a L } with L co-authors. The notations used in this article are summarized in Table 1.
In the name disambiguation task, i often represents an author name instead of a real-life person. This means d 1 i and d 2 i may not be written by the same author. Our task is to find the correct partition function such that each partition only contains documents related to one person p. Given a dataset D i , we hope to group the documents into K disjoint clusters C i = { C 1 i , C 2 i , , C K i } where the number of clusters is unique for each author i. K is the number of real people that share the same author name. The partition function is:
Φ ( D i ) C i
where C k i = { d j i | Φ ( d j i ) = p k , d j i D i }, denoting the paper cluster of a real-life person p k , k { 1 , 2 , , K } .

3.2. Mutual Information and Information Bottleneck

Given random variables (X, Y), mutual information I ( X ; Y ) is a measurement of the mutual dependence between the two variables calculating by their marginal probability distributions p ( X ) , p ( Y ) and joint distribution p ( X , Y ) :
I ( X ; Y ) = x , y p ( x , y ) log p ( x , y ) p ( x ) p ( y )
The entropy of variable X, denoted by H ( X ) , standing for uncertainty for random variables is defined as:
H ( X ) = x p ( x ) log p ( x ) = I ( X ; X )
Let x X be the inputs of a neural network and y Y be corresponding labels, t T are the output of hidden layers. In general neural networks, Y X T forms a Markov chain. Information Bottleneck principle is proposed to provide a technique for extracting information from the input random variables, which is related to predicting some output random variables. In other words, Representation T is a compression of X and a prediction of Y. In learning progress, neural network models need to learn an efficient representation T of X while preserve prediction capability of Y, which could be formulated by the following optimization function:
min p ( t | x ) , Y X T { I ( X ; T ) β I ( T ; Y ) }
where β is a tradeoff parameter between compression and prediction accuracy. In DNNs and CNNs, IB principle claims that deep neural networks consist of two stages of fitting and compression.

4. Framework

In this section, we present details of our name disambiguation model. We first introduce document representation and variational Graph Auto-Encoder, which mainly focus on encoding attributes of each paper. The similarity metric is calculated by the relationship between these attributes. Then we introduce Graph Embedding to our model to leverage the relationships between authors.
As shown in Figure 1, the colorful dots represent document representations. We use a word2vec model to calculate these representation vectors. Then a variational Graph Auto-Encoder and a Graph Embedding model are applied to fine-tune these vectors. After representation learning, we use Hierarchical Agglomerative Clustering to partition documents into different clusters.

4.1. Feature Embedding

Recent works to learn the representation of words mainly focus on neural network-based methods. Among these works, word2vec models [30] including Continuous Bag-of-Words (CBOW) and Skip-gram model could extract the relationships between words and perform better than other works.
Given a series of training words w 1 , w 2 , , w T , CBOW learns word representation based on the co-occurrence of words within the size of a predefined context window in the training corpus. The model is based on the idea that the probability of occurrence of words around a word can predict what the word is. It maximizes the co-occurrence probability between words that appear in the predefined context window. The objective is to maximize the probability:
1 T t = 1 T c j c , j 0 log p ( w t | w t + j )
in which c is the context window and the loss function is optimized by a neural network. After training, the weight matrix is used to represent the training words.
A specific document d n could be represented as { f 1 , f 2 , , f M } and f denotes the stemmed words of the title, co-authors, orgnization and venue. One hot vector is utilized to stands for each feature f. We use CBOW model here to compress each document representation into a low-dimensional vector. After training the neural network, the m-th weight f ^ m of the neural network weight matrix is the low-dimensional representation related to feature f m . Every document representation d n could be calculated by d n = i = 1 M α i f ^ i . α i is the inverted document frequency of f i to reduce the weight of some irrelevant stemmed words like prepositions. D i = { d 1 , d 2 , d N } captures the similarity between these documents using the co-occurrence probability of each paper’s features.

4.2. Variational Graph Auto-Encoder

Recall D i is the set of document representation related to the author name i. We remove the superscript i in the notations for simplicity and the implication of the symbol is unchanged. For each author name, we could build a graph G = ( D , E ) where D = { d 1 , d 2 , d n } is the document representation output D i defined in Section 4.1. The document embedding vectors could represent graph nodes and E is the adjacency matrix A to represent edges between these nodes.
The adjacency matrix A is constructed by two approaches: one is calculating affiliation similarity between two documents, and the other one is using a collection of high-precision features.
We measure the similarity between two papers by the weighted sum of feature representation. If the similarity between document d i and d j is larger than the threshold, we construct an edge between these two nodes and A i j = A j i = 1 . We empirically set this threshold to 25. For other node pairs, the relevant value in the adjacency matrix is 0. Due to the lack of labeled data, we also use some high precision rules to bootstrap positive pairs. For example, there is little chance that two authors with the same name work in the same organization and have co-authors work together. Besides, one’s corporators are relatively fixed in a period of time. If two publications are written by exactly the same co-authors, an edge is also constructed between them.
Variational Graph Auto-Encoder (as shown in Figure 2) is used in our model to improve the generalization ability of the model. Let X = [ d 1 T , d 2 T , ] be the representation matrix of documents associated with an author name. The encoder is a two-layer graph convolutional network:
q ( Z | X , A ) = i = 1 N q ( z i | X , A )
with
q ( z i | X , A ) = N ( z i | μ i , diag ( σ i 2 ) )
μ = [ μ 1 , μ 2 , ] = G C N μ ( X , A ) is mean value matrix and log σ = [ log σ 1 , log σ 2 , ] = G C N σ ( X , A ) represents variance matrix. The two-layer GCN is computed by
G C N ( X , A ) = A ˜ R e l u ( A ˜ X W 0 ) W 1
where W 0 is identical in G C N μ ( X , A ) and G C N σ ( X , A ) while W 1 is different, A ˜ = D 1 2 A D 1 2 and D is degree matrix. The decoder reconstructs the adjacency matrix A ^ :
p ( A | Z ) = i = 1 N j = 1 N p ( A i j | z i , z j )
with
p ( A i j | z i , z j ) = sigmoid ( z i T z j )
The goal is to make the reconstruction matrix as close as possible to the original one. The loss function is as follows:
L = E q ( Z | X , A ) [ log p ( A | Z ) ] KL [ q ( Z | X , A ) | | p ( Z ) ]
The first item refers to cross entropy loss. In the second item, p ( Z ) = p ( z i ) , p ( z i ) = N ( 0 , 1 ) is standard normal distribution. KL [ q ( · ) | | p ( · ) ] is the Kullback-Leibler divergence which denotes the similarity between distribution p and q.
We take the latent matrix output of encoder Z = [ z 1 T , z 2 T , ] as new embedding vectors of these documents. By knowing the relationship of existing nodes, the new matrix is a higher-level expression and have the predictability of new node pairs.

4.3. Graph Embedding

In the Graph Auto-Encoder, we only consider the relationship of co-authors and affiliations between two documents. As demonstrated in Table 2, we cannot determine whether document 1 and document 2 are written by the same author “Yanjun Zhang” by measuring feature similarity. If we have a look at document 3, we could know that one-hop collaborators of document 1 and document 2 are completely overlapped, so the “Yanjun Zhang” in document 1 and document 2 is the same real-life person. In order to utilize relationships of co-authors and papers, the intuition is to represent a graph network as a matrix.
There are several ways to extract graph topology information, like deepwalk [31] and graph convolutional network [32]. Inspired by previous work [5], we try to make related authors and papers closer utilizing author-author, author-document, and document-document graphs. Different from their model, we wish to initialize the author and document node representation with prior knowledge. We preserve the ranking orders based on letting the embedding vector of a document closer to all related items (including documents and their authors).
Given an author name i, we construct a person-person network G p p = ( A , E p p ) refering to the collaborators’ relationship within all documents related to author i. A is the set of all co-authors related to author name i and e i j E p p is the edge between author a i and author a j . The weight of edge e i j is the number of collaborations between two authors and the value is 0 if they never co-published a paper.
Similarly, a document-document network G d d = ( D , E d d ) reflects the graph topology between the set of documents associated with author i. E d d considers one-hop co-authors. If two documents have more than one one-hop collaborator, they are likely to be similar and we build an edge between them. D is constructed by the output matrix Z of Graph Auto-Encoder such that we include feature engineering into Graph Embedding.
More importantly, we would like to introduce author relationship to guide the document clustering. G p d = ( A D , E p d ) is the bipartite network that connect documents and authors. A i is the set of collaborators of all documents in D i . The edge is defined as the number of author a i appears in the document d j and the value is 1 if the paper is written by this author and otherwise 0. If two papers have a large overlap of co-authors set (their authors are close to each other in the author-author graph topology), they are considered to be similar as well.
The similarity score between two documents is defined as S i j d d = d i T d j . Let document i and document j be positive pairs, we would like to maximize the S i j = d i T d j such that the cosine similarity is smaller. We find node pairs based on the ranking orders and the probability of preserving the orders is defined as:
p ( S i j > S i t | d i , d j , d t ) = sigmoid ( S i j d d S i t d d )
For a set of documents D , the probability is:
p ( D ) = ( d i , d j ) P G d d ( d i , d t ) N G d d sigmoid ( S i j d d S i t d d )
Aiming to fine-tune the document embedding vectors, we need to maximize the probability. The loss function of the linked document network is:
L d d = min D ln p ( D )
In the same way, we could derive the loss function for preserving the ranking orders of the author-author graph and bipartite network:
L p p = min A ln p ( A )
L p d = min A , D ln p ( A , D )
For this three Graph Embedding model, our goal is to embed these three graphs, so the optimization function is:
L = min A , D L d d + L p p + L p d + λ ( | | A | | F 2 + | Z | | F 2 )
where A and Z are co-author matrix and document matrix.
After training the Graph Embedding model, we combine the topology information of three graphs as well as feature similarity into the document embedding matrix. Let the output embedding matrix be Y = [ y 1 T , y 2 T , ] and for every author name the matrix is different.

4.4. Clustering

We apply Hierarchical Agglomerative Clustering (HAC) [33] afterward. This algorithm treats each data point in the training sample set as a cluster at first. Then it calculates the distance between every two groups and merges the two most similar clusters that are closest until the number of clusters equal to the number of true clusters.

5. Calculating Mutual Information

Researchers plot information plane by calculating mutual information I ( X ; T ) and I ( T ; Y ) . The calculation of mutual information in classification tasks have been studied. In this section, we will show how we do calculations step by step in a clustering problem.
As our proposed method consists of several parts, we evaluate I ( X ; T ) and I ( T ; Y ) separately for each component. In our task, input X stands for the original representation matrix. The original document representation is calculated by the CBOW model. Therefore, X = D = { d 1 , d 2 , , d N } . And Y stands for true clustering labels.
Tishby [28] utilized toy data for demonstration and separated the neuron’s output into 30 bins between 1 and 1. In practical deep neural networks with real datasets, T is usually a high-dimensional matrix. Thus, in our calculation, a softmax function is applied to the output of each component for squashing a C-dimensional vector z to a C-dimensional vector with values in the range [ 0 , 1 ] :
sigmoid ( t ) j = e t j i = 1 C e t i
Now we get output of neurons ranging from 0 to 1 and then we also bin each neuron’s output into 30 bins range from 0 to 1. The output of this discretization progress is representation T.
Cheng [29] noted that the network’s parameters Θ are fixed after training, which means p ( t i | x i ) = 1 . t i could be represented as t i = f ( x i , Θ ) . In this case, H ( T | X ) = 0 . The mutual information between X and T could be calculated by:
I ( X ; T ) = H ( T ) H ( T | X ) = H ( T ) = t p ( t ) log p ( t )
Representation T contains information about input X and thus the calculation takes this information into account. During our calculation, the network’s parameters are also fixed. Therefore, we use the above formula to calculate I ( X ; T ) . In Feature Embedding, representation T is exactly X and mutual information can be calculated by I ( X ; T ) = I ( X ; X ) = H ( X ) .
Different from image classification tasks, embedding matrix T of documents is not directly associated with labels in our model. We use HAC to establish links between documents. Therefore, I ( T ; Y ) is estimated by the mutual information between the output of HAC and labels. Thus, in our approach I ( X ; T ) represents the relation between the outputs of each component and original document embedding vectors. I ( T ; Y ) is another measurement of clustering results.

6. Experiments

6.1. Datasets

Many existing name disambiguation methods are validated by no more than 20 author datasets. We evaluate our proposed approach on a dataset of online academic search and data mining system AMiner [3]. This system automatically extracts and storages researchers’ personal files from the web. It also integrates academic data from other online databases such as DBLP, ACM Digital Library, CiteSeer, and SCI. For a fair comparison, we validate our approach with the same sampled dataset used in [4] consisting of 100 author names. Some methods require external information (e.g., social linkage) that is hard to access and not provided by popular open datasets. Therefore, these models are not compared in our experiments.

6.2. Competing Method

We validate the performance of our model by comparing it with several state-of-the-art graph-based models as following:
Zhang and Yao [4]: The method proposes a name disambiguation method by incorporating both global and local information. In global supervision, human annotation is required to generate triplet sampling, while in the local context, a Graph Auto-Encoder is used.
Zhang et al. [5]: This method constructs three local graphs within an author name according to author and paper similarity. By optimizing the pairwise ranking objective, the graph topology is embedded into a low-dimensional matrix.
GHOST [6]: GHOST is a model based on author acquaintance. Every author’s name associated with papers represents a node in the GHOST graph. By selecting valid paths, it computes the similarity between two nodes and uses the AP algorithm for clustering.
Union finding: Union finding is to build graph linkage based on several strictly matching of co-authors and affiliations. A clustering is constructed by all nodes that have a corporate relationship.

6.3. Experimental Results

We evaluate the results with pairwise precision, recall and F1-score and calculate macro score of each metric for all 100 names. For a fair comparison, we select the same 15 name references as in [4] and compare these three metrics for each name reference.
We use the same hyperparameters for each dataset. In the CBOW model, the dimension of document representation vectors is set to 100 and the context window is set to 5. In the variational Graph Auto-Encoder, the threshold of inverse document frequency is 25, the output dimension of the first layer in graph convolution network is 200, the output dimension of the second layer is set to 100, the learning rate is 0.01 and the model trained for 200 epochs. In the graph network embedding model, the learning rate is 0.05 and the regularization parameter is 0.01 .
Table 3 shows that the average F1-score of our methods is + 3.87 % higher than Zhang and Yao, + 25 % over Zhang et al., and + 33.85 % over GHOST. It performs the best for 11 (out of 15) name references.
Our model achieves the best performance comparing with other graph-based methods in F1-score. Further, we would like to know whether changes in precision or recall lead to the variation in F1-measure. Comparing pairwise precision between our solution and other methods, we could see that the average precision of our approach is a little higher than that of Zhang and Yao, Zhang et al. and the Union finding, but lower than GHOST. Moreover, the average recall of our clustering results is much higher than the approach Zhang and Yao, Zhang et al. and GHOST, but lower than Union Finding. High recall value demonstrates that our model is capable to find more documents that belong to the right categories, thus the proportion of positive pairs in all true pairs increases. Although Union finding achieves the highest recall, this method would divide documents that do not belong to author i into the cluster of his documents and make the wrong predictions, leading to much lower precision. Similarly, GHOST is more conservative by ignoring many documents which belong to the author. Overall, our method performs the best.
GHOST constructs a similarity matrix according to the co-author graph. Zhang et al. leverage the graph topology by relying on a co-author graph and the paper-paper graph. Zhang and Yao utilizes global supervision and local linkage paper-paper graph. Some methods focus on the relationship between two documents, and others concentrate on graph topology among the set of all authors. By combining the advantages of both and making the best of the information, our model outperforms these baseline methods, even if we do not use labeled data as Zhang and Yao did to generate triplet samplings.
Figure 3a,b are true cluster visualization of embedding spaces while Figure 3c,d denote low dimensional space of predict clusters. Different categories are distinguished by different colors. The F1-score of our approach on this set is 0.6338 while the method of Zhang and Yao only achieves 0.5382 . From the circles with dashed lines in Figure 3b,d we could see that samples in a true cluster are clustered into different parts in prediction because of the spatial dissimilarity of representation vectors using their model. Our approach does better clustering by allowing similar vectors to be closer. The orange scattered points shown in Figure 3a indicate that these representation vectors are closer in the embedding space by our approach. These scattered points are classified correctly in Figure 3c.

6.4. Component Analysis

Each component in our approach is evaluated separately. Feature Embedding represents using only document representation to measure similarity. Graph Auto-Encoder and Graph Embedding are constructed on previous Feature Embedding. As shown in Table 4, compared with Feature Embedding, Graph Auto-Encoder and Graph Embedding improve the performance by 6.41 % and 4.83 % respectively. Our method is superior to the above sub-systems as the precision and recall are both the highest.
In addition, we evaluated each component by reducing the dimension of embedding vectors associated with a publication set. As shown in Figure 4, we use true labels distribution to reflect the impact of each sub-system on the embedding space. As we can see, Graph Auto-Encoder brings the green dots as well as the blue dots together while Graph Embedding lets these points be closer so that they can be clustered more accurately. Outlier orange dots, which are circled by grey dash lines, could also be found and moved to the correct area. This means that our model is effective for outliers.

6.5. Evaluating by Information Theory

We calculate I ( X ; T ) and I ( T ; Y ) on a candidate set for each component respectively and make more explanations of our model. X is the original document representation after feature embedding, T is the output of each component and Y stands for the matrix of true clustering labels.
From Table 5, we observe that I ( X ; T ) is 5.3230 for Feature Embedding and after one epoch training of Graph Auto-Encoder the value drops to 3.8771 , which suggests that introducing linkage function by Graph Auto-Encoder weakens the relationship between the new embedding matrix and the original one. Meanwhile I ( T ; Y ) instantly increases to 2.2304 . Also, we could see that the trends of I ( T ; Y ) and F1-score are exactly the same. Both of them can reflect the quality of the clustering results.
Researchers found DNNs and CNNs could be divided into two phases of extracting information and compression via information plane. We wish to know whether our graph networks still follows this two-stage process. Although during experiments, our graph networks do not work this way. we could still utilize a similar method to find out what role each component plays and how it affects the final result.
Figure 5a,b are mutual information paths through training Graph Auto-Encoder. In Figure 5a, I ( X ; T ) keeps increasing during the training progress, indicating that training Graph Auto-Encoder could learn more information from X. Meanwhile, mutual information I ( T ; Y ) tends to be stable and increases a little (equals to 2.2781 after 200 epochs, see in Table 5). Combining with the fact that I ( T ; Y ) instantly increases after one epoch training (shown in Table 5), we could infer that the improvement of clustering results is mainly based on the structure itself. As we introduce the linkage function of paper relationships with Graph Convolutional Network in GAE, other GCNs are worth trying for solving this problem. Besides, we could do less training without damaging the performance.
Figure 6 shows the mutual information of I ( X ; T ) and I ( T ; Y ) when training Graph Embedding. We observe that I ( X ; T ) keeps decreasing during training progress, which is contrary to the case in Graph Auto-Encoder. This is mainly due to the intrinsic property of the Graph Embedding that continuously introduces additional information in the process of training. In our experiment, Graph Embedding supplements T with additional information of person-person, document-document, person-document graph topology, leading to weaker and weaker dependence between X and T, and thus I ( X ; T ) keeps decreasing during training. Similar to the scenario of Graph Auto-Encoder, I ( T ; Y ) only slightly increases during the training progress (from 2.2659 to 2.2801 , see in Table 5), coinciding with the small improvement of F1-score (from 0.7918 to 0.8155 , see in Table 5). During the actual training progress, I ( T ; Y ) and F1-score increases during the first several training epochs, but decreases after training too many rounds. The reason for this is that I ( X ; T ) is too small, suggesting that we lose too much information from the original representation matrix. As the original representation vectors are constructed by attributes, there is a tradeoff between information of relationships and attributes. As we train these components separately, the model could not incorporate the information from both attributes and graph topology at the same time. Thus, one model incorporating relationships and attributes may be better for clustering.
In Information Bottleneck principle, I ( X ; T ) decreases as hidden layers keep compressing information from input X while in our model the decrease suggests introducing additional information. Although the analysis results are not the same, but the analysis method utilizing mutual information is consistent and gives us a better understanding of these models.

7. Conclusions

In this paper, we introduce a method to address the problem of author name disambiguation. Our proposed representation model is capable of encoding attributes and graph topology of all authors and papers by incorporating a Graph Auto-Encoder and a Graph Embedding model. We evaluate our approach on a sampled dataset of an online academic search and data mining system which integrates data from popular databases. Experimental results show that our approach outperforms several state-of-the-art graph-based methods. In addition, we increase the interpretability of each component through information theory and show that it is capable of guiding the model selection and training progress in future work.

Author Contributions

Y.M. proposes the main idea, performs the experiments and write the manuscript. Y.W. reviews and edits the manuscript. C.L. investigates related work and proposes some idea. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by NSFC (No. 61901267).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, J.; Berzins, K.; Hicks, D.; Melkers, J.; Xiao, F.; Pinheiro, D. A boosted-trees method for name disambiguation. Scientometrics 2012, 93, 391–411. [Google Scholar] [CrossRef]
  2. Tran, H.N.; Huynh, T.; Do, T. Author name disambiguation by using deep neural network. In Asian Conference on Intelligent Information and Database Systems; Springer: Berlin/Heidelberg, Germany, 2014; pp. 123–132. [Google Scholar]
  3. Tang, J.; Zhang, J.; Yao, L.; Li, J.; Zhang, L.; Su, Z. Arnetminer: Extraction and mining of academic social networks. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 990–998. [Google Scholar]
  4. Zhang, Y.; Zhang, F.; Yao, P.; Tang, J. Name Disambiguation in AMiner: Clustering, Maintenance, and Human in the Loop. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, London, UK, 19–23 August 2018; pp. 1002–1011. [Google Scholar]
  5. Zhang, B.; Al Hasan, M. Name disambiguation in anonymized graphs using network embedding. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 1239–1248. [Google Scholar]
  6. Fan, X.; Wang, J.; Pu, X.; Zhou, L.; Lv, B. On graph-based name disambiguation. J. Data Inf. Qual. (Jdiq) 2011, 2, 10. [Google Scholar] [CrossRef]
  7. Cetoli, A.; Akbari, M.; Bragaglia, S.; O’Harney, A.D.; Sloan, M. Named Entity Disambiguation using Deep Learning on Graphs. arXiv 2018, arXiv:1810.09164. [Google Scholar]
  8. Kim, K.; Rohatgi, S.; Giles, C.L. Hybrid Deep Pairwise Classification for Author Name Disambiguation. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–7 November 2019; pp. 2369–2372. [Google Scholar]
  9. Bunescu, R.; Paşca, M. Using encyclopedic knowledge for named entity disambiguation. In Proceedings of the 11th conference of the European Chapter of the Association for Computational Linguistics, Trento, Italy, 4 April 2006. [Google Scholar]
  10. Han, H.; Giles, L.; Zha, H.; Li, C.; Tsioutsiouliklis, K. Two supervised learning approaches for name disambiguation in author citations. In Proceedings of the 2004 Joint ACM/IEEE Conference on Digital Libraries, Tucson, AZ, USA, 7–11 June 2004; pp. 296–305. [Google Scholar]
  11. Delgado, A.D.; Martínez, R.; Montalvo, S.; Fresno, V. Person name disambiguation on the web in a multilingual context. Inf. Sci. 2018, 465, 373–387. [Google Scholar] [CrossRef]
  12. Emami, H. A graph-based approach to person name disambiguation in Web. Acm Trans. Manag. Inf. Syst. (Tmis) 2019, 10, 1–25. [Google Scholar] [CrossRef]
  13. Cen, L.; Dragut, E.C.; Si, L.; Ouzzani, M. Author disambiguation by hierarchical agglomerative clustering with adaptive stopping criterion. In Proceedings of the 36th International ACM SIGIR Conference on Research and Development in Information Retrieval, Dublin, Ireland, 28 July–1 August 2013; pp. 741–744. [Google Scholar]
  14. Wu, H.; Li, B.; Pei, Y.; He, J. Unsupervised author disambiguation using Dempster–Shafer theory. Scientometrics 2014, 101, 1955–1972. [Google Scholar] [CrossRef]
  15. Song, Y.; Huang, J.; Councill, I.G.; Li, J.; Giles, C.L. Efficient topic-based unsupervised name disambiguation. In Proceedings of the 7th ACM/IEEE-CS Joint Conference on Digital Libraries, Vancouver, BC, Canada, 18–23 June 2007; pp. 342–351. [Google Scholar]
  16. Tang, J.; Fong, A.C.; Wang, B.; Zhang, J. A unified probabilistic framework for name disambiguation in digital library. IEEE Trans. Knowl. Data Eng. 2011, 24, 975–987. [Google Scholar] [CrossRef]
  17. Tang, J.; Qu, M.; Mei, Q. Pte: Predictive text embedding through large-scale heterogeneous text networks. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; pp. 1165–1174. [Google Scholar]
  18. Levin, M.; Krawczyk, S.; Bethard, S.; Jurafsky, D. Citation-based bootstrapping for large-scale author disambiguation. J. Am. Soc. Inf. Sci. Technol. 2012, 63, 1030–1047. [Google Scholar] [CrossRef] [Green Version]
  19. Louppe, G.; Al-Natsheh, H.T.; Susik, M.; Maguire, E.J. Ethnicity sensitive author disambiguation using semi-supervised learning. In Proceedings of the International Conference on Knowledge Engineering and the Semantic Web, Prague, Czech Republic, 21–23 September 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 272–287. [Google Scholar]
  20. Giles, C.L.; Zha, H.; Han, H. Name disambiguation in author citations using a k-way spectral clustering method. In Proceedings of the 5th ACM/IEEE-CS Joint Conference on Digital Libraries (JCDL’05), Denver, CO, USA, 7–11 June 2005; pp. 334–343. [Google Scholar]
  21. Hussain, I.; Asghar, S. DISC: Disambiguating homonyms using graph structural clustering. J. Inf. Sci. 2018, 44, 830–847. [Google Scholar] [CrossRef]
  22. Km, P.; Mondal, S.; Chandra, J. A Graph Combination With Edge Pruning-Based Approach for Author Name Disambiguation. J. Assoc. Inf. Sci. Technol. 2020, 71, 69–83. [Google Scholar] [CrossRef]
  23. Zhang, W.; Yan, Z.; Zheng, Y. Author Name Disambiguation Using Graph Node Embedding Method. In Proceedings of the 2019 IEEE 23rd International Conference on Computer Supported Cooperative Work in Design (CSCWD), Porto, Portugal, 6–8 May 2019; pp. 410–415. [Google Scholar]
  24. Wang, X.; Tang, J.; Cheng, H.; Philip, S.Y. Adana: Active name disambiguation. In Proceedings of the 2011 IEEE 11th International Conference on Data Mining, Vancouver, BC, Canada, 11–14 December 2011; pp. 794–803. [Google Scholar]
  25. Shin, D.; Kim, T.; Choi, J.; Kim, J. Author name disambiguation using a graph model with node splitting and merging based on bibliographic information. Scientometrics 2014, 100, 15–50. [Google Scholar] [CrossRef]
  26. Delgado, A.D.; Martínez, R.; Fresno, V.; Montalvo, S. A data driven approach for person name disambiguation in web search results. In Proceedings of the COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, Dublin, Ireland, 23–29 August 2014; pp. 301–310. [Google Scholar]
  27. Saxe, A.M.; Bansal, Y.; Dapello, J.; Advani, M.; Kolchinsky, A.; Tracey, B.D.; Cox, D.D. On the information bottleneck theory of deep learning. J. Stat. Mech. Theory Exp. 2019, 2019, 124020. [Google Scholar] [CrossRef]
  28. Shwartz-Ziv, R.; Tishby, N. Opening the black box of deep neural networks via information. arXiv 2017, arXiv:1703.00810. [Google Scholar]
  29. Cheng, H.; Lian, D.; Gao, S.; Geng, Y. Utilizing Information Bottleneck to Evaluate the Capability of Deep Neural Networks for Image Classification. Entropy 2019, 21, 456. [Google Scholar] [CrossRef] [Green Version]
  30. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 3111–3119. [Google Scholar]
  31. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  32. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  33. Ward, J.H., Jr. Hierarchical grouping to optimize an objective function. J. Am. Stat. Assoc. 1963, 58, 236–244. [Google Scholar] [CrossRef]
Figure 1. This figure shows the framework of our model. After constructing representation vectors by a word2vec model, we do representation learning through a Graph Auto-Encoder and a Graph Embedding model. Then we cluster the documents using their representation vectors.
Figure 1. This figure shows the framework of our model. After constructing representation vectors by a word2vec model, we do representation learning through a Graph Auto-Encoder and a Graph Embedding model. Then we cluster the documents using their representation vectors.
Entropy 22 00416 g001
Figure 2. Variational Graph Auto-Encoder.
Figure 2. Variational Graph Auto-Encoder.
Entropy 22 00416 g002
Figure 3. Visualization of clustering results of our proposed method and method of Zhang and Yao. We compress the representation vector space into a 2-dimensional space on a candidate set. Different colors in (a,b) represent different true clusters while different colors in (c,d) denote different predicted clusters.
Figure 3. Visualization of clustering results of our proposed method and method of Zhang and Yao. We compress the representation vector space into a 2-dimensional space on a candidate set. Different colors in (a,b) represent different true clusters while different colors in (c,d) denote different predicted clusters.
Entropy 22 00416 g003
Figure 4. Visualization of clustering results of each component. Different colors denote different true clusters. Dashed grey ellipses indicate document representations in a true cluster are moved closer and closer by our method.
Figure 4. Visualization of clustering results of each component. Different colors denote different true clusters. Dashed grey ellipses indicate document representations in a true cluster are moved closer and closer by our method.
Entropy 22 00416 g004
Figure 5. Mutual information paths of Graph Auto-Encoder.
Figure 5. Mutual information paths of Graph Auto-Encoder.
Entropy 22 00416 g005
Figure 6. Mutual information paths of Graph Embedding.
Figure 6. Mutual information paths of Graph Embedding.
Entropy 22 00416 g006
Table 1. Notations.
Table 1. Notations.
NameDescription
D the set of all documents included in the database
| D | the number of documents in the database
iauthor name i
D i the document set associated with the author name i
d n i document n associated with the author name i
fa stemmed word of an attribute
preal-life person p
A i the set of all collaborators of author i
C i clusters for author i
Table 2. Related documents according to one-hop co-author.
Table 2. Related documents according to one-hop co-author.
Document 1:
Author: Yanjun Zhang, Jun Shan, Zhanjun Liu, Fanqin Li
Orgnization: College of Chemical Engineering, Department of Applied Chemistry,
Document 2:
Author: Yanjun Zhang, Guanghan Zuo, Jun Ji
Orgnization: College of Chemical Engineering Ctr. For Element and Strct. Anal.
Document 3:
Author: Jun Shan, Zhanjun Liu, Guanghan Zuo, Jun Ji
Orgnization: Department of Applied Chemistry, Ctr. For Element and Strct. Anal.
Table 3. Clustering Results of Different Name Disambiguation Methods. The F1-scores highlighted in bold show that our approach performs the best for 11 (out of 15) name references.
Table 3. Clustering Results of Different Name Disambiguation Methods. The F1-scores highlighted in bold show that our approach performs the best for 11 (out of 15) name references.
NameOur ApproachZhang and YaoZhang et al.GHOSTUnion Finding
PrecRecF1PrecRecF1PrecRecF1PrecRecF1PrecRecF1
Xu Xu70.7370.0670.3974.1845.8656.6847.7339.9843.5161.3421.7932.157.2266.2913.02
Rong Yu73.0742.8153.9989.1346.5161.1266.5336.947.4792.0036.4152.1716.0041.2223.05
Yong Tian78.1261.0368.5376.3251.9561.8273.1856.3463.6686.9454.5867.0610.7894.5019.36
Lu Han60.0235.2944.4551.7828.0536.3946.0517.9525.8369.7217.3927.8415.1096.3026.10
Lin Huang74.7250.4360.2177.1032.8746.0969.4333.1344.8686.1517.2528.745.9133.7610.06
Kexin Xu91.0989.6590.3791.3798.6494.8785.7444.1358.2792.9028.5243.6477.6383.6280.51
Wei Quan77.2348.8659.8553.8839.0245.2674.4133.9446.6286.4227.8042.0737.1696.5753.67
Tao Deng76.3853.8263.1581.6343.6256.8655.2527.9337.1173.3324.5036.7312.5564.7621.03
Hongbin Li70.0978.4874.0577.2069.2172.9965.7952.8658.6256.2929.1238.3912.9294.5922.73
Hua Bai65.8641.9451.2571.4939.7351.0854.9335.9743.4783.0629.5443.5822.0893.2335.71
Meiling Chen79.6047.8459.7674.9344.7055.9979.2225.1538.1886.1123.8537.3524.8366.9236.22
Yanqing Wang49.0066.4956.4271.5275.3373.3772.7342.6253.7480.7940.3953.8624.1266.9535.46
Xudong Zhang74.5522.2434.2662.4022.5433.1255.638.1114.1685.757.2313.3465.1247.3654.84
Qiang Shi53.3850.3651.8352.2036.1542.7243.3337.9940.4953.7226.8035.7618.1186.3729.94
Min Zheng59.8520.7630.8257.6522.3532.2153.6217.6326.5480.5015.2125.5811.9574.4820.60
Average78.1067.4772.4077.9663.0369.7070.2248.7257.5381.7240.4354.0940.7876.5253.20
Table 4. Clustering Results of Each Component.
Table 4. Clustering Results of Each Component.
PrecRecF1
Feature Embedding72.2950.1459.21
Graphh Auto-Encoder75.5358.0165.62
Graph Embedding77.7154.4664.04
Overall78.1067.4772.40
Table 5. Mutual Information and F1-score for Each Component.
Table 5. Mutual Information and F1-score for Each Component.
I ( X ; T ) I ( T ; Y ) F1
Feature Embedding5.32302.03790.6209
Graph Auto-Encoder (after one epoch training)3.87712.23040.7834
Graph Auto-Encoder (after training)4.51862.27810.7939
Graph Embedding (after one epoch training)4.36352.26590.7918
Graph Embedding (after training)1.51752.28010.8155

Share and Cite

MDPI and ACS Style

Ma, Y.; Wu, Y.; Lu, C. A Graph-Based Author Name Disambiguation Method and Analysis via Information Theory. Entropy 2020, 22, 416. https://doi.org/10.3390/e22040416

AMA Style

Ma Y, Wu Y, Lu C. A Graph-Based Author Name Disambiguation Method and Analysis via Information Theory. Entropy. 2020; 22(4):416. https://doi.org/10.3390/e22040416

Chicago/Turabian Style

Ma, Yingying, Youlong Wu, and Chengqiang Lu. 2020. "A Graph-Based Author Name Disambiguation Method and Analysis via Information Theory" Entropy 22, no. 4: 416. https://doi.org/10.3390/e22040416

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop