Next Article in Journal
Federated Learning for Medical Imaging Segmentation via Dynamic Aggregation on Non-IID Data Silos
Next Article in Special Issue
Energy-Constrained UAV Data Acquisition in Wireless Sensor Networks with the Age of Information
Previous Article in Journal
Deep Learning for Forecasting-Based Applications in Cyber–Physical Microgrids: Recent Advances and Future Directions
Previous Article in Special Issue
Deep Learning Recommendations of E-Education Based on Clustering and Sequence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Transformer-Based User Alignment Model across Social Networks

1
Institute of Information Technology, Information Engineering University, Zhengzhou 450002, China
2
Songshan Laboratory, Zhengzhou 450018, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(7), 1686; https://doi.org/10.3390/electronics12071686
Submission received: 15 March 2023 / Revised: 28 March 2023 / Accepted: 31 March 2023 / Published: 3 April 2023
(This article belongs to the Special Issue AI Technologies and Smart City)

Abstract

:
Cross-social network user identification refers to finding users with the same identity in multiple social networks, which is widely used in the cross-network recommendation, link prediction, personality recommendation, and data mining. At present, the traditional method is to obtain network structure information from neighboring nodes through graph convolution, and embed social networks into the low-dimensional vector space. However, as the network depth increases, the effect of the model will decrease. Therefore, in order to better obtain the network embedding representation, a Transformer-based user alignment model (TUAM) across social networks is proposed. This model converts the node information and network structure information from the graph data form into sequence data through a specific encoding method. Then, it inputs the data to the proposed model to learn the low-dimensional vector representation of the user. Finally, it maps the two social networks to the same feature space for alignment. Experiments on real datasets show that compared with GAT, TUAM improved ACC@10 indicators by 11.61% and 16.53% on Facebook–Twitter and Weibo–Douban datasets, respectively. This illustrates that the proposed model has a better performance compared to other user alignment models.

1. Introduction

With the development and popularization of the Internet and related mobile devices, social networks became an important part of people’s lives. A large number of people share their lives, work, or exchange information with each other on multiple social networking platforms to meet their different social requirements. For example, a user can either share their daily life on Facebook or express their opinion on Twitter. Although multiple social networks greatly enrich people’s lives, there are many problems involving the joint research and analysis of multiple networks, and the problem of user alignment across social networks came into being. This problem aims to combine multiple social networks to analyze the relationships between nodes and construct a high-quality cross-social network user alignment model to connect the same person on different social networks. This model can be widely used in various fields, such as cross-network recommendation [1], cross-domain information diffusion [2,3], link prediction [4], and network dynamics analysis [5].
For user alignment across social networks, some researchers proposed a user alignment method based on user location information. Riederer et al. [6] utilized the rich user position information in location-based social networks (LBSNs) to propose a POIS algorithm, which started from the user’s trajectory data, analyzed the similarity of user pairs, and designed a general and self-tunable algorithm to align users between two LBSNs. Chen et al. [7] used kernel density estimation (KDE) to alleviate the data sparsity in measuring user similarity, and further organized location data based on the structure of the grid. Then, they pruned and reduced the search space to improve the efficiency of user alignment.
At the same time, some researchers proposed a large number of cross-social network user alignment methods based on user profile information [8,9,10,11]. This information includes a variety of profile information about users in two networks, such as usernames, educational experiences, cities of residence, and personal descriptions. It showed that the cross-social network user alignment model based on user profile information is feasible and effective for some social networks. Zhang et al. [11] proposed MOBIUS, where user similarity between different social networks is measured by extracting some user characteristics, such as prefixes, suffixes, the rarity of usernames, and user habits. Zhao et al. [12] proposed a BP neural network mapping for social network alignment, which used the BP neural network to obtain the mapping between user name vectors across social networks, changed the classification problem into a mapping problem between vectors, and improved the accuracy of social network alignment. However, in the actual social network scenario, the user’s profile information is difficult to obtain, which involves the user’s privacy. Many user profile information cannot be accessed, and many users will imitate others or forge personal information for various purposes, making the user alignment method based on user profile information difficult to work.
Aiming at the user alignment model based on user profiles, some researchers believed that the same person has a similar structure in different social networks, so they proposed cross-social network user alignment methods based on the user’s local social structure [13,14], and user-based local and global social structure [15]. However, the actual situation is that due to the different service functions of different social networks, the same person has different social network structures in different social networks. In this regard, some researchers applied network embedding representations to cross-social network user alignment. Feng et al. [16] proposed a hypergraph neural network (HGNN) framework for data representation learning, which encodes higher-order data correlations in the hypergraph structure, and a hyperedge convolution operation to process these correlations achieves good results. On this basis, Chen et al. [17] proposed a multi-layer graph convolutional network (MGCN) that jointly considers the local network structure and hypergraph structure. In addition, a two-stage spatial coordination mechanism is proposed to efficiently align users across different large-scale social networks. Although user alignment models based on network embedding were proven to be effective, the problem of “too close” representation of network embedding is also unavoidable, which greatly affects the accuracy of the model. Yan et al. [18] introduced pseudo-anchors to make the distribution of user embedding representations more uniform and proposed a meta-learning algorithm to guide the update of pseudo-anchors, which effectively solved the problem that the network embedding representations are too close. In sparse networks, user network structural similarities are small and difficult to identify. Li et al. [19] proposed a triple-layer attention mechanism-based network embedding (TANE) method, which learns latent structural information by using the weighted structural similarity of the first-order and second-order neighbors to reduce network sparsity, and fully mines the network structure to identify users. He et al. [20] proposed a heuristic algorithm based on the attention mechanism HDyNA, which obtained the local importance weight of new nodes in a single network through the attention mechanism. It used the anchor node as supervision information and heuristically learned the local influence driven by the alignment task of new nodes to improve the performance of model alignment across dynamic networks. To reduce the expression of noise edges for structural consistency across social networks, Liu et al. [21] proposed a network structure denoising framework, which learned the user network topology and removes noise edges by iterative learning through a parameter sharing encoder and graph neural network (GNN) to improve the structural similarity across networks. Zheng et al. [22] considered the influence of distribution differences between different networks on model performance, and a periodically consistent adversarial mapping model (CAMU) was proposed, which learned the mapping function across potential representation spaces and solved the representation distribution difference through adversarial training between the mapping function and discriminator. In addition, periodic consistency training can alleviate the overfitting problem and reduce the number of labeled users required. The comparisons of the user alignment models are listed in Table 1. Most of the existing models did not assign weight and some of them assigned local weight and had the problem of “over-smooth”.
However, the above methods based on GNN to mine user network structure information will appear “over-smooth” with the deepening of the number of network layers. That is to say, the characteristics of all nodes in the same connected component tend to be consistent after multiple convolution operations, resulting in an extreme decrease in the effect of the model. Inspired by Yin et al. [23], graph structure information is encoded into the model via Transformer. To fill the research gaps, a Transformer-based user alignment model (TUAM) across social networks is proposed in this paper, which accurately imports the graph structure information into the model through three encoding methods, calculates the semantic similarity between cross-social network nodes, and obtains the accurate expression of network nodes to solve the problem of “over-smooth”.
The main conclusions and novelties of this paper can be summarized as follows: First, a Transformer-based user alignment model (TUAM) is proposed to model node embeddings in social networks. This method transforms the graph structure data into a sequence data type that is convenient for Transformer learning through three novel graph structure encoding methods, which effectively avoids the phenomenon of “over-smooth” of GNN. Second, TUAM can assign the weight of different users’ influence and network structure, accurately model the embedding vector of users, and improve the accuracy of social network alignment. Third, experiments on real datasets Facebook–Twitter and Weibo–Douban show that the results of the proposed model are superior to existing models.

2. Methodology

2.1. Development of Transformer

With the advent of better computer hardware, such as graphic processing units (GPUs), and word embedding methods, such as Word2Vec and Glove, deep learning models, such as convolutional neural networks (CNN) and recurrent neural networks (RNN) gained wider use in building natural language processing (NLP) systems. However, word-by-word processing of RNN limits computational efficiency, so Vaswani et al. [24] proposed a deep learning model Transformer based on self-attention, which contains layers of stacked encoders and decoders that allow it to learn complex linguistic information. In the field of NLP, the ability of Transformer and self-supervised learning is combined to develop a Transformer-based pre-training language model (T-PTLM). A generative pre-trained Transformer (GPT) is based on T-PTLM and developed at the Transformer decoder layer. BERT [25] is the first T-PTLM developed based on the Transformer encoder layer. The study by Kaplan et al. [26] showed that the performance of T-PTLM could be improved simply by increasing the size of the model, and the results drive the large-scale development of T-PTLM, such as GPT-3 [27], PANGU [28], and GShard [29]. Some of which can contain billions of parameters, and switch-Transformers [30] contain trillions of parameters.
With the success of T-PTLM, they are also used in other fields, such as finance, biomedicine, computer vision, etc. Dosovitskiy et al. [31] proposed the Transformer-based ViT model, which is simple, effective, and extensible, and became a milestone work in the application of Transformer in the field of computer vision. However, the computational complexity of the self-attention module of the ViT model is very high. Therefore, Liu et al. [32] proposed the Swin Transformer model, which not only adopts a pyramidal hierarchical structure, but also proposes a linear complexity attention calculation, which is very powerful in downstream tasks.

2.2. Transformer Model

Transformer is a deep learning model based entirely on the self-attention mechanism, which replaces long short-term memory (LSTM) with the attention mechanism. Transformer abandons the inherent mode of the previous traditional encoder–decoder model that must be combined with CNN or RNN, which not only reduces the computation complexity and improves parallel efficiency, but also is higher in accuracy and performance than the popular RNN.
The Transformer consists of two parts: encoders and decoders. Each layer of the encoder has two sublayers, one is the multi-head attention mechanism, and the other is the position fully connected feed-forward network. Each sublayer uses a residual connection and layer normalization. Unlike encoders, decoders insert a third sublayer in addition to two sublayers in the encoder layer, which performs a multi-head attention mechanism learning model on the output of the encoder.
The inputs of the self-attention module are represented as X = [ x 1 T , , x N T ] R N × d , where d is the dimension of input features. The corresponding Q, K, and V can be calculated using the input matrix X and the linear array matrix W Q R d × d q , W K R d × d k , and W V R d × d v , where, d q , d k , and d v are the corresponding feature dimensions, respectively. Assume that d = d q = d k = d v , and they can be calculated as [24]
Q = X W Q , K = X W K , V = X W V .
The output of the self-attention module is calculated as follows:
A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d k ) V .

2.3. Problem Definition

Cross-social network user alignment is also known as user identity linkage [33]. It is different from predicting the connection relationship between two or more different users on a single network [34,35,36,37]. Instead, it is to find correspondence between different identities of the same user in multiple social networks. In this section, we first introduce some necessary definitions and then give a formal definition of user alignment across social networks.
Definition 1.
Social networks: Represented as  G = ( V , E , X ) , where  V = { v i | i = 1 , , N }  is a set of user nodes,  E = { e i j = ( v i , v j ) | v i V , v j V }  is a set of user edges,  e i j = ( v i , v j )  represents the connection status of node i and node j. If node i has a connection with node j,  e i j = 1 , otherwise  e i j = 0 .  X = { x i | i = 1 , , N }  denotes a set of user feature vectors,  x i  is the feature vector of the ith user.
Definition 2.
Source social networks and target social networks: the problems of cross-social network user alignment between two social networks are mainly studied, so the two social networks are named the source social networks  G S  and the target social networks  G T , where the source social network is  G S = ( V S , E S , X S )  and the target social network is  G T = ( V T , E T , X T ) .
Definition 3.
Anchor user: Given source social network as  G S = ( V S , E S )  and target social network as  G T = ( V T , E T ) , where the user set belonging to the same person is the anchor user set   T = { ( u s , v t ) | u s V S , v t V T }  and  e i j = ( u i s , v j t )  represents the anchor link between  G S  and  G T . The cross-social network user alignment problem is essentially equivalent to the anchor link prediction problem between  G S  and  G T .

3. A Transformer-Based User Alignment Model

The proposed TUAM includes three encoding design methods to aim at the “over-smooth” problem in the current cross-social network user alignment problem based on GNN. The overall framework of the proposed model is shown in Figure 1.
The proposed model represents the node embeddings of different social networks to the same vector space, and finally obtains the alignment results according to the node embedding vector similarity matrix.

3.1. Centrality Encoding

Node centrality is used to measure the importance of nodes in the graph and it is important information in the graph structure, such as celebrities who are followed by everyone is an important factor in predicting the trend of social networks [38,39]. However, this information is often ignored in previous graph convolution operations and is very valuable.
In the proposed model, centrality encoding takes degree centrality, which is one of the standard centrality measures, as an additional signal to the model, and assigns each node two real-valued embedding vectors based on its in-degree and out-degree. Since each node is centrally encoded, it can be added to the node feature as the input of the proposed model. The important information of nodes is input to the model through centrality encoding, and the semantic correlation and node importance between nodes are obtained through the attention mechanism. Centrality encoding formulas is [23]
h i ( 0 ) = x i + z deg ( v i ) + z deg + ( v i ) + ,
where z , z + R d are the corresponding learnable embedding vectors with dimension d in directed graphs of in-degree deg ( v i ) and out-degree deg + ( v i ) , respectively. In undirected graphs, in-degree deg ( v i ) and out-degree deg + ( v i ) can be unified to deg ( v i ) .

3.2. Spatial Encoding

One advantage of the Transformer model over GNN is its global receptive field. In the Transformer layer, the attention mechanism can focus on and process information at any location. Notably, the position dependencies in individual node locations or the encoding layer need to be clear. Therefore, for sequence data, there are two ways to indicate node position information. One is to assign an absolute position, that is, absolute position encoding, and the other is to encode the relative distance of any two nodes in the Transformer layer, that is, relative position encoding. However, nodes are not sequential in graphs, they can be in multi-dimensional space and connected by edges. To obtain the structural information of the coding diagram in the model, this paper adopts a novel spatial encoding method. To measure the spatial relationship between two nodes v i and v j in the graph, a function ϕ ( v i , v j ) : V × V R is designed to represent the shortest path distance between v i and v j . If there is no connection relationship between the two nodes, a special value is set for each element in the similarity matrix A as a learnable scalar as a bias term in the self-attention module, the formula is as follows [23]:
A i j = ( h i W Q ) ( h j W K ) T d + b ϕ ( v i , v j ) ,
where b ϕ ( v i , v j ) is a learnable scalar indexed by ϕ ( v i , v j ) and is shared across all layers.
Compared to traditional GNN, where the receptive field of GNN is confined to the neighborhood, the Transformer layer provides a global receptivity field, and each node can follow all other nodes in the graph. At the same time, nodes in the Transformer layer can adjust their attention to all other nodes through b ϕ ( v i , v j ) .

3.3. Edge Encoding

In graphs, the structural features of edges are also important for graph representation learning, and encoding them into the network is essential. To better encode edge features into the attention layer, the shortest path S P i j = ( e 1 , e 2 , , e N ) from v i to v j for each node pair can be found, and the average of the edge feature embedding representation is calculated to incorporate the edge feature into the attention module as a bias term. The edge encoding formula is [23]
c i j = 1 N n = 1 N x e n ( w n E ) T ,
where x e n is the feature of the nth edge e N in S P i j , w n E R d E is the embedded weight matrix for the nth edge, and d E is the number of edge feature dimensions.
Therefore, the output after three encodings is [23]
A i j = ( h i W Q ) ( h j W K ) T d + b ϕ ( v i , v j ) + c i j .
After obtaining the similarity matrix A, the output Y from the attention module can be calculated, and the calculation formula is as follows:
Y = s o f t m a x ( A ) V = s o f t m a x ( Q K T d + b ϕ + C ) V .

3.4. User Identification Layer

Through the above method, the user representation matrices Y S and Y T from the source social network and the target social network can be obtained. The reciprocal of the Euclidean distance L 2 is used to measure the similarity matrix S between each user vector, and the alignment results are obtained according to the similarity. The calculation formula is
S i j = ( y s i y t j 2 + ε ) 1 , y s i Y S , y t j Y T ,
where ε is a special minimum value.
Different from the traditional cross-social network alignment model, the cross-entropy of any pair of nodes ( v i , v j ) , v i Y S , v j Y T is not used as the loss function [17] in this paper, because the positive and negative sample size of the association between the two networks is large. Therefore, the method of maximizing the probability of the positive side and minimizing the probability of the negative side as the loss function is adopted. On the one hand, the nodes with link relationships between the two networks are more similar. On the other hand, the distribution of nodes without connection relationships will be more scattered. The calculation formula is
L = mean ( l j i ( v i , v j ) O p o s log η ( S i j ) ) + mean ( ( 1 l j i ) ( v i , v j ) O n e g log ( 1 η ( S i j ) ) ) ,
where η is the sigmoid function. The value of v s i G S and v t j G T depends on whether and belongs to the set of anchor users T. The formula of l j i is
l j i = { 1 ,   ( v s i , v t j ) T 0 ,   ( v s i , v t j ) T .

4. Datasets and Experiments

4.1. Datasets

The datasets used in the experiment are two real-world datasets from Cao [40]: Facebook–Twitter and Weibo–Douban. Both datasets are collected from public information in the social network, so there is no privacy breach. Table 2 lists this data and some basic information.
Facebook–Twitter: Facebook and Twitter are both social networking platforms with large numbers of users, and the dataset was collected through third-party platforms dedicated to linking users, collecting a total of 1,107,695 accounts, of which 422,291 accounts were related to Facebook, 669,198 accounts were related to Twitter, and 328,224 pairs of users were associated between the two datasets.
Weibo–Douban: Weibo and Douban are China’s largest microblogging sites and movie rating sites, respectively. For the Douban dataset, there are 1,694,399 active users, of which 141,614 are associated with Weibo users. In addition to having social relationships in Weibo’s network with Douban, the dataset also collects user-generated content, such as Douban’s movie rating history and Weibo’s blog history. The average user has 287 blogs and 120 rating histories. Here, we only use the information of social connections.
In Table 2, |V| is the number of users, |E| is the number of edges, and |CV| is the number of associated users in the two social networks.

4.2. Comparable Models

DeepWalk: uses random walks to sample the node sequence, and then uses the word2vec model to learn the node embedding.
GCN: extracts features from graph data through GNN for node representation learning, which is widely used in node classification, graph classification, and link prediction.
HGCN: bases on hypergraph convolutional networks for network embedding.
MGCN: combines graph convolutional networks (GCN) and hypergraph convolutional networks to jointly learn network vertex representations at different levels of granularity.
GAT: introduces an attention mechanism based on graph convolution to the weighted summing of the features of neighboring nodes and learning node representation.

4.3. Evaluation Metric

To evaluate the performance of the models, we use the most commonly used evaluation metric: accuracy@K (ACC@K), which is defined as
A C C @ K = 1 N i = 0 N 1 A v i ( @ k ) ,
where A v i ( @ k ) indicates whether the source social network user v i ( v i Y S ) corresponds to the user in the target social network v i ( v i Y T ) exists in the top k users, and N is the total number of test users in the source social network. In addition, A v i ( @ k ) is defined as
A v i ( @ k ) = { 1 ,   v i t o p k 0 ,   v i t o p k .

4.4. Analysis of Experimental Results

Table 3 and Table 4 show the experimental results on the Facebook–Twitter and Weibo–Douban datasets, showing that the proposed model outperforms the comparable methods on both datasets.
In this table, the ACC@K, K = 1, 10, 20, and 50 of the existing and proposed user alignment models on the Facebook–Twitter dataset are compared. The results show that the proposed TUAM performs better than other models.
According to Table 3 and Table 4, it is illustrated that the TUAM model outperforms other comparable models in the two datasets. When K equals to 10 or 20, the performance of TUAM has the most significant improvement compared to other models. When K = 10, the accuracy rate ACC@10 in Facebook–Twitter and Weibo–Douban improved by 11.61% and 16.53% compared to GAT, respectively. The DeepWalk model embeds nodes by random walk, but this process does not map the node features of the two networks to the same vector space, but directly performs user alignment, which severely reduces the accuracy of the model. GCN model can effectively aggregate the surrounding neighbor information, but it does not consider the importance of each neighbor node, which limits the learning ability of the features of network nodes. HGCN and MGCN are based on GCN, which greatly benefits from the nonlinearity of neural networks, but the modeling of hyperedge is very redundant for non-hypergraph problems, which not only has no benefit to the model, but also reduces the accuracy of the model. The attention mechanism in GAT perfectly solves the problem of GCN, making GAT perform better than GCN in the two datasets, but it is still limited by the “over-smooth” problem and cannot mine the characteristics of network nodes at a deeper level. However, the three encodings of TUAM can learn network structure information as well as GNN, or even better, and will not be limited by the “over-smooth” problem, can mine network node features at a deeper level, and TUAM’s global receptive field makes the model learn higher-level structural features better, so it performs better than other models. The accuracy rates of these models on two datasets are shown in Figure 2 and Figure 3.
In Figure 2 and Figure 3, it is obvious that the accuracy of TUAM on two datasets is higher than other existing models, especially when K = 50. The ablation experiments on the importance of the three encodings are conducted in TUAM across the Facebook–Twitter and the Weibo–Douban datasets. The ablation results are shown in Table 4 and Table 5. The best results are indicated in bold font.
According to Table 5 and Table 6, the experimental results show that for the TUAM without centrality coding, spatial coding, or edge coding, the effectiveness of the TUAM decreases. This is because the centrality encoding module can effectively encode the information of different nodes into Transformers to improve the accuracy of model recognition. At the same time, the spatial encoding module and edge encoding can effectively capture the spatial information and structural information of nodes, which is more conducive to the expression of the structural characteristics of the Transformer learning network. The three kinds of encoding methods can effectively convert the graph topology information into sequence data information that is conducive to Transformer learning. TUAM does not need to consider the “over-smooth” phenomenon caused by too-deep layers, such as GCN, and has a larger global receptive field, efficiently learns topology structure features, and improves model performance. The trends of accuracy on Facebook–Twitter and Weibo–Douban datasets of ablation experiments are shown in Figure 4 and Figure 5.
It is shown that TUMA with three encoding methods has better performance on accuracy on two datasets compared to the model without one of the three encodings.

5. Conclusions

In this paper, user alignments across social networks were described. The research received a lot of attention in both academia and industry, and was involved in many social network-related applications, such as link prediction, interest recommendation, etc. A Transformer-based user alignment model based on network topology information was proposed to learn the structural information between nodes in the networks. The proposed model is different from the traditional GCN through graph convolution to obtain network structure information from neighbor nodes, but through a specific encoding method to express the graph structure information in the form of sequence data. Experiment results show that the proposed method can better describe the association relationship between node neighbors, has a more accurate vector representation of nodes, and can improve the accuracy of user association matching.
While our approach has certain advantages, there are also some drawbacks. Our work only makes use of the structural information of the network, which is less informative. If additional attribute information is considered, it will be helpful to improve the performance of the model. At the same time, the proposed model cannot be adapted to large-scale graph datasets. Therefore, the future research direction is to build a framework for integrating social network structure information and attribute information on the basis of this user alignment model. In addition, it is necessary to reduce the cost of model calculation to make it suitable for large-scale alignment across social networks.

Author Contributions

Conceptualization, L.J. and S.L.; methodology, G.W. and F.P.; software, L.W. and T.L.; validation, T.L. and S.L.; investigation, T.L., L.W., and G.W.; writing—original draft preparation, T.L.; writing—review and editing, G.W. and F.P.; funding acquisition, L.J. and S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Songshan Laboratory Project, Major Science and Technology Project of Henan province, No: 221100210700-2.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, S.; Zhang, J.; Wang, L.; Hua, X.S. Social friend recommendation based on multiple network correlation. IEEE Trans. Multimed. 2016, 18, 287–299. [Google Scholar] [CrossRef] [Green Version]
  2. Peng, C.; Xu, K.; Wang, F.; Wang, H. Predicting information diffusion initiated from multiple sources in online social networks. In Proceedings of the 6th International Symposium on Computational Intelligence and Design, Washington, DC, USA, 28–29 October 2013; Volume 2, pp. 96–99. [Google Scholar]
  3. Zafarani, R.; Liu, H. Users joining multiple sites: Distributions and patterns. In Proceedings of the 8th International Conference on Weblogs and Social Media, Michigan, MI, USA, 1–4 June 2014; pp. 635–638. [Google Scholar]
  4. Zhang, J.; Yu, P.S.; Zhou, Z.H. Meta-path based multi-network collective link prediction. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 1286–1295. [Google Scholar]
  5. Zafarani, R.; Liu, H. Users joining multiple sites: Friendship and popularity variations across sites. Inf. Fusion 2016, 28, 83–89. [Google Scholar] [CrossRef]
  6. Riederer, C.; Kim, Y.; Chaintreau, A.; Korula, N.; Lattanzi, S. Linking users across domains with location data: Theory and validation. In Proceedings of the 25th International Conference on World Wide Web, Montréal QC, Canada, 11–15 April 2016; pp. 707–719. [Google Scholar]
  7. Chen, W.; Yin, H.; Wang, W.; Zhao, L.; Zhou, X. Effective and efficient user account linkage across location based social networks. In Proceedings of the 34th IEEE International Conference on Data Engineering (ICDE), Paris, France, 16–19 April 2018; pp. 1085–1096. [Google Scholar]
  8. Liu, J.; Zhang, F.; Song, X.; Song, Y.I.; Lin, C.Y.; Hon, H.W. What’s in a name? An unsupervised approach to link users across communities. In Proceedings of the 6th ACM International Conference on Web Search and Data Mining, Rome, Italy, 4–8 February 2013; pp. 495–504. [Google Scholar]
  9. Liu, S.; Wang, S.; Zhu, F.; Zhang, J.; Krishnan, R. HYDRA: Large-scale social identity linkage via heterogeneous behavior modeling. In Proceedings of the 2014 ACM SIGMOD International Conference on Management of Data, Snowbird Utah, UT, USA, 22–27 June 2014; pp. 51–62. [Google Scholar]
  10. Tan, S.; Guan, Z.; Cai, D.; Qin, X.; Bu, J.; Chen, C. Mapping users across networks by manifold alignment on hypergraph. In Proceedings of the 28th AAAI Conference on Artificial Intelligence, Québec, QC, Canada, 27–31 July 2014; pp. 159–165. [Google Scholar]
  11. Zhang, H.; Kan, M.Y.; Liu, Y.; Ma, S. Online social network profile linkage. In Proceedings of the 10th Asia Information Retrieval Societies Conference, Kuching, Malaysia, 3–5 December 2014; pp. 197–208. [Google Scholar]
  12. Zhao, Y.; Liu, Y.; Guo, X.; Sun, X.; Wang, S. User naming conventions mapping learning for social network alignment. In Proceedings of the 2021 13th International Conference on Computer and Automation Engineering (ICCAE), Melbourne, Australia, 20–22 March 2021; pp. 36–42. [Google Scholar]
  13. Zhou, X.; Liang, X.; Zhang, H.; Ma, Y. Cross-platform identification of anonymous identical users in multiple social media networks. IEEE Trans. Knowl. Data Eng. 2016, 28, 411–424. [Google Scholar] [CrossRef]
  14. Liu, L.; Cheung, W.K.; Li, X.; Liao, L. Aligning users across social networks using network embedding. In Proceedings of the 25th International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; pp. 1774–1780. [Google Scholar]
  15. Zhang, W.; Shu, K.; Liu, H.; Wang, Y. Graph neural networks for user identity linkage. arXiv 2023, arXiv:1903.02174. [Google Scholar]
  16. Feng, Y.; You, H.; Zhang, Z.; Ji, R.; Gao, Y. Hypergraph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27–31 January 2019; pp. 3558–3565. [Google Scholar]
  17. Chen, H.; Yin, H.; Sun, X.; Chen, T.; Gabrys, B.; Musial, K. Multi-level graph convolutional networks for cross-platform anchor link prediction. In Proceedings of the 26th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual, CA, USA, 23–27 August 2020; pp. 1503–1511. [Google Scholar]
  18. Yan, Z.; Liu, L.; Li, X.; Cheung, W.K.; Zhang, Y.; Liu, Q.; Wang, G. Towards improving embedding based models of social network alignment via pseudo anchors. IEEE Trans. Knowl. Data Eng. 2021, 35, 4307–4320. [Google Scholar] [CrossRef]
  19. Li, Y.; Cui, H.; Liu, H.; Li, X. Triple-layer attention mechanism-based network embedding approach for anchor link identification across social networks. Neural. Comput. Appl. 2022, 34, 2811–2829. [Google Scholar] [CrossRef]
  20. He, J.; Liu, L.; Yan, Z.; Wang, Z.; Xiao, M.; Zhang, Y. User alignment across dynamic social networks based on heuristic algorithm. In Proceedings of the 7th International Conference on Systems and Informatics (ICSAI), Chongqing, China, 13–15 November 2021; pp. 1–7. [Google Scholar]
  21. Liu, L.; Wang, C.; Zhang, Y.; Wang, Y.; Liu, Q.; Wang, G. Denoise network structure for user alignment across networks via graph structure learning. In Proceedings of the 7th Data Mining and Big Data, Beijing, China, 21–24 November 2022; pp. 105–119. [Google Scholar]
  22. Zheng, C.; Pan, L.; Wu, P. CAMU: Cycle-consistent adversarial mapping model for user alignment across social networks. IEEE Trans. Cybern. 2022, 52, 10709–10720. [Google Scholar] [CrossRef] [PubMed]
  23. Ying, C.; Cai, T.; Luo, S.; Zheng, S.; Ke, G.; He, D.; Shen, Y.; Liu, T.Y. Do Transformers really perform badly for graph representation? In Proceedings of the Annual Conference on Neural Information Processing Systems 2021, Virtual, 6–14 December 2021; pp. 28877–28888. [Google Scholar]
  24. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomes, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  25. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  26. Kaplan, J.; Mccandlish, S.; Henighan, T.; Brown, T.B.; Chess, B.; Child, R.; Gray, S.; Radford, A.; Wu, J.; Amodei, D. Scaling Laws for Neural Language Models. arXiv 2023, arXiv:2001.08361. [Google Scholar]
  27. Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020; pp. 1877–1901. [Google Scholar]
  28. Zeng, W.; Ren, X.; Su, T.; Wang, H.; Liao, Y.; Wang, Z.; Jiang, X.; Yang, Z.; Wang, K.; Zhang, X.; et al. Pangu: Large-scale autoregressive pretrained Chinese language models with auto-parallel computation. arXiv 2023, arXiv:2104.12369. [Google Scholar]
  29. Lepikhin, D.; Lee, H.; Xu, Y.; Chen, D.; Firat, O.; Huang, Y.; Krikun, M.; Shazeer, N.; Chen, Z. Gshard: Scaling giant models with conditional computation and automatic sharding. arXiv 2023, arXiv:2006.16668. [Google Scholar]
  30. Fedus, W.; Zoph, B.; Shazeer, N. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. arXiv 2023, arXiv:2101.03961. [Google Scholar]
  31. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Unterthiner, T.; Dehghani, M.; Heigold, G.; Gelly, S.; Uszkoreit, J.; Houlsby, N. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations, Virtual, Austria, 3–7 May 2021. [Google Scholar]
  32. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar]
  33. Mu, X.; Zhu, F.; Lim, E.P.; Xiao, J.; Wang, J.; Zhou, Z.H. User identity linkage by latent user space modelling. In Proceedings of the 22nd ACM SIGKDD International Conference Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1775–1784. [Google Scholar]
  34. Huang, Z.; Lin, D.K.J. The time-series link prediction problem with applications in communication surveillance. Inf. J. Comput. 2009, 21, 286–303. [Google Scholar] [CrossRef] [Green Version]
  35. Lee, C.; Pham, M.; Jeongm, M.K.; Kim, D.; Lin, D.K.J.; Chavalitwongse, W.A. A network structural approach to the link prediction problem. Inf. J. Comput. 2015, 27, 249–267. [Google Scholar] [CrossRef]
  36. Li, Z.; Fang, X.; Bai, X.; Sheng, O.R.L. Utility-based link recommendation for online social networks. Manag. Sci. 2015, 63, 1657–2048. [Google Scholar] [CrossRef] [Green Version]
  37. Yang, J.; Zhang, X.D. Predicting missing links in complex networks based on common neighbors and distance. Sci. Rep. 2016, 6, 38208. [Google Scholar] [CrossRef]
  38. Marshall, P.D. The promotion and presentation of the self: Celebrity as marker of presentational media. Celebr. Stud. 2010, 1, 35–48. [Google Scholar] [CrossRef] [Green Version]
  39. Marwick, A.; Boyd, D. To see and be seen: Celebrity practice on Twitter. Convergence 2011, 17, 139–158. [Google Scholar] [CrossRef] [Green Version]
  40. Cao, X.; Yu, Y. ASNets: A benchmark dataset of aligned social networks for cross-platform user modeling. In Proceedings of the 25th ACM International Conference on Information and Knowledge Management, Indianapolis, IN, USA, 24–28 October 2016; pp. 1881–1884. [Google Scholar]
Figure 1. The overall framework of TUMA across social networks.
Figure 1. The overall framework of TUMA across social networks.
Electronics 12 01686 g001
Figure 2. The accuracy of several models on the Facebook–Twitter dataset.
Figure 2. The accuracy of several models on the Facebook–Twitter dataset.
Electronics 12 01686 g002
Figure 3. The accuracy of several models on the Weibo–Douban dataset.
Figure 3. The accuracy of several models on the Weibo–Douban dataset.
Electronics 12 01686 g003
Figure 4. The accuracy of ablation experiments on the Facebook–Twitter dataset.
Figure 4. The accuracy of ablation experiments on the Facebook–Twitter dataset.
Electronics 12 01686 g004
Figure 5. The accuracy of ablation experiments on the Weibo–Douban dataset.
Figure 5. The accuracy of ablation experiments on the Weibo–Douban dataset.
Electronics 12 01686 g005
Table 1. The comparisons of the existing user alignment models.
Table 1. The comparisons of the existing user alignment models.
MethodsReferencesUser LocationUser ProfileNetwork StructureWeight AssignmentOpen DatasetsOver-Smooth
LocalGlobal
Non-deep learning[6]---
[7]---
[11]---
Deep learning[12]--
[16]--
[17]--
[18]--
[19]--
[20]--
[21]--
[22]--
TUAM--
Table 2. The information of datasets.
Table 2. The information of datasets.
Dataset|V||E||CV|
Facebook348172241874
Twitter32116020
Weibo 112411625471
Douban 211701695
1 http://www.weibo.com/ (accessed on 31 March 2023), 2 http://www.douban.com/ (accessed on 31 March 2023).
Table 3. The performance of different methods on ACC@K on the Facebook–Twitter dataset.
Table 3. The performance of different methods on ACC@K on the Facebook–Twitter dataset.
DatasetModelACC@1ACC@10ACC@20ACC@50
Facebook–TwitterDeepWalk0%1.32%1.97%6.58%
GCN3.22%31.11%40.00%53.33%
HGCN0.89%9.43%18.87%47.17%
MGCN1.42%10.32%20.15%58.30%
GAT3.77%28.89%42.22%62.22%
TUAM8.01%40.50%57.85%78.51%
Table 4. The performance of different methods on ACC@K on the Weibo–Douban dataset.
Table 4. The performance of different methods on ACC@K on the Weibo–Douban dataset.
DatasetModelACC@1ACC@10ACC@20ACC@50
Weibo–DoubanDeepWalk0.21%2.34%9.55%14.23%
GCN9.77%39.92%49.89%62.85%
HGCN4.25%30.15%47.56%59.45%
MGCN5.31%31.85%50.32%63.69%
GAT11.25%37.37%52.65%66.88%
TUAM12.06%53.90%65.25%90.07%
Table 5. Ablation experiment results on the Facebook–Twitter dataset.
Table 5. Ablation experiment results on the Facebook–Twitter dataset.
Centrality EncodingSpatial EncodingEdge EncodingACC@1ACC@10ACC@20ACC@50
×6.94%35.41%48.40%72.16%
×7.65%32.56%46.62%68.86%
×6.58%32.92%47.86%68.51%
8.01%40.50%57.85%78.51%
Table 6. Ablation experiment results on the Weibo–Douban dataset.
Table 6. Ablation experiment results on the Weibo–Douban dataset.
Centrality EncodingSpatial EncodingEdge EncodingACC@1ACC@10ACC@20ACC@50
×7.09%51.77%67.38%85.80%
×9.22%48.94%63.12%83.69%
×6.38%41.84%63.83%78.72%
12.06%53.90%65.25%90.07%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lei, T.; Ji, L.; Wang, G.; Liu, S.; Wu, L.; Pan, F. Transformer-Based User Alignment Model across Social Networks. Electronics 2023, 12, 1686. https://doi.org/10.3390/electronics12071686

AMA Style

Lei T, Ji L, Wang G, Liu S, Wu L, Pan F. Transformer-Based User Alignment Model across Social Networks. Electronics. 2023; 12(7):1686. https://doi.org/10.3390/electronics12071686

Chicago/Turabian Style

Lei, Tianliang, Lixin Ji, Gengrun Wang, Shuxin Liu, Lan Wu, and Fei Pan. 2023. "Transformer-Based User Alignment Model across Social Networks" Electronics 12, no. 7: 1686. https://doi.org/10.3390/electronics12071686

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop