Next Article in Journal
Polarities of Exceptional Geometries of Type E6
Previous Article in Journal
Analysis of Competitive Strategies and Cost Effects in the Digitalization of the Shipping Industry
Previous Article in Special Issue
Representative Community Detection Algorithms for Attribute Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Link Prediction Algorithm Based on Layer Attention Mechanism for Multiplex Networks

1
School of Information Science and Engineering, Shenyang University of Technology, Shenyang 110870, China
2
Shenyang Key Laboratory of Advanced Computing and Application Innovation, Shenyang 110870, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(23), 3803; https://doi.org/10.3390/math13233803
Submission received: 30 August 2025 / Revised: 1 October 2025 / Accepted: 26 November 2025 / Published: 27 November 2025
(This article belongs to the Special Issue Deep Representation Learning for Social Network Analysis)

Abstract

Link prediction is a technique for predicting future or missing relationships between entities based on current network information. Recent studies on link prediction in multiplex networks have improved prediction accuracy by considering inter-layer similarity, but they do not fully exploit the distinct topological information present in each individual network layer. Therefore, this paper proposes a link prediction algorithm based on a layer attention mechanism for multiplex networks, called LATGCN, which simultaneously considers both inter-layer and intra-layer information. Firstly, it uses two-layer GCN to capture the intra-layer embedding representation of a multiplex network. Secondly, it employs the layer attention mechanism to find the layer importance score of each layer of the multiplex network, and utilizes the layer importance scores to weight the intra-layer information according to the difference in importance. Thirdly, we design an embedding representation fusion module to integrate the intra-layer embedding representation and global embedding representation. Finally, experimental results on four real-world social networks and two biological networks show that LATGCN outperforms all compared methods across four evaluation metrics, demonstrating its effectiveness for link prediction.

1. Introduction

Link prediction aims to exploit the underlying topological structure of a network to infer potential or missing relationships between nodes [1]. With the rapid development of information, many social software programs have emerged [2]. In many real-world networks, multiple types of relationships can exist simultaneously between the same pair of nodes. For example, person A and person B may be friends on both WeChat and Weibo, while having no connection on Twitter [3]. Such networks, consisting of multiple single-layer networks, are referred to as multiplex networks, where each layer shares the same set of nodes but differs in edge distribution. There are many applications of link prediction in multiplex networks. In social networks, link prediction can be used to infer unknown links based on existing edges, thereby recommending potential friends [4]. In protein interaction networks, it can predict interactions between proteins [5]. Similarly, in traffic networks, link prediction can help forecast traffic trends [6].
In recent years, the research on link prediction for multiplex networks has received much attention and many methods have been gradually proposed [7]. The research mainly relies on three frameworks, i.e., a framework based on inter-layer similarity and node similarity, framework based on network representation learning, and hybrid framework based on inter-layer similarity and network representation learning. However, multiplex network link prediction becomes more challenging when considering the complex inter-layer relationships and topologies of multiplex networks. To more accurately predict the relationships in multiplex networks, some methods [8,9] were proposed by considering the complex inter-layer relationships and topologies. However, they failed to fully utilize the rich inter-layer relationships in multiplex networks, which resulted in the differences in the contribution of information from other layers to the target layer and poor model performance. Recent hybrid methods, such as the hybrid framework LISCNE [10], which belongs to network representation learning, aim to extract more comprehensive structural information by incorporating the importance of inter-layer information. LISCNE assigns higher weights to layers more similar to the target layer and lower weights to less similar layers. However, the method still has limitations. It simply obtained the common vector by aggregating the graphs of all layers. When edge distributions differ significantly across layers, this aggregation method struggles to capture such differences.
To accurately extract the embedding representation for each layer of the network and determine inter-layer similarity, we propose a link prediction algorithm based on layer attention mechanism for multiplex networks, called LATGCN, which simultaneously addresses the above limitations and enhances the accuracy of link prediction. Firstly, an embedding representation extraction module is proposed to efficiently generate low-dimensional node embeddings, which are capable of effectively capturing the topological structure information of each layer in the multiplex network. Secondly, to improve the accuracy of link prediction, a layer attention mechanism module is designed, enabling the model to more effectively exploit inter-layer relationships and enhance the comprehensiveness of the embedding representations. Thirdly, we design an embedding representation fusion module which integrates the intra-layer embedding representations and global embedding representation to obtain comprehensive and high-quality multiplex network topology information. Finally, the comprehensive experiments are conducted on six publicly available realistic network datasets. Our method is compared to representative and state-of-the-art methods on four metrics. The experimental results confirm the superiority of LATGCN.
The rest of the paper is organized as follows. Section 2 briefly describes related work. Section 3 first describes the multiplex network link prediction problem and then describes our proposed model in detail. Experimental results and analysis are shown in Section 4. Section 5 summarizes the paper and outlines our future work.

2. Related Work

In this section, we review existing work related to link prediction, including methods for single-layer networks as well as approaches specifically designed for multiplex networks.

2.1. Link Prediction Methods for Single-Layer Networks

Many researchers have tackled the problem of link prediction in single-layer networks by considering the similarity between nodes. Common Neighbors (CN) [11] was the classical similarity metric which considered the number of common neighbors of two nodes. It considered that the more common neighbors there were, the higher the similarity between two nodes. Since then, many new metrics have been proposed inspired by CN, such as Salton [12], Sorenson [13], and HPI [14]. Adamic-Adar (AA) [15] differentiated the contributions of different neighbors by assigning weights. Based on AA, Zhou et al. [16] proposed Resource Allocation, which assumes that common neighbors with lower degree contribute more to node similarity. Lü et al. [17] further proposed Local Path, which considered local path information and was based on the idea that two nodes were more likely to form a link if multiple shorter paths existed between them. Aziz [18] integrated information on common neighbors, node degrees, and the distribution of edges connected to the target node. However, the above methods only considered the local information of the network and neglected the global topology. To address this limitation, subsequent researchers proposed Katz [19], which considered the global path information. The Katz metric measured the contribution of all possible paths between two nodes, assigning higher weights to shorter paths. However, its computational complexity was high, making it unsuitable for large-scale networks.
Some other researchers have approached link prediction in single-layer networks by considering node representation learning. DeepWalk [20], also known as a graph embedding algorithm, was inspired by Word2vec [21] in natural language processing. DeepWalk treated random walk sequences in a network as sentences and nodes as words, thereby learning node representations in a continuous vector space analogous to word embeddings. Subsequently, Grover et al. [22] extended DeepWalk by introducing biased random walk strategies that incorporated both depth-first and breadth-first sampling mechanisms, leading to the development of Node2vec. Methods such as GraRep [23] and SDNE [24], which preserve higher-order neighborhood information, were then proposed. The convolution operator in Convolutional Neural Network [25] inspired the development of Graph Neural Network [26]. Researchers extended the convolutional operator to network representation learning and proposed Graph Convolutional Network (GCN) [27]. Graph Generative Adversarial Network [28] combined generative and discriminative models with network representations to learn low-dimensional vector representations of nodes. To address the challenge of learning embeddings for only some nodes in large-scale networks, SepNE [29] was proposed, which independently learned the representations of different subsets of nodes during network separation. Yao et al. [30] proposed the MCAS, which introduced multi-scale subgraphs as input graphs. It obtained complementary network structures from different perspectives and used contrastive learning to achieve information balance among multi-scale subgraphs. Su et al. [31] proposed the VGNNDP, a new approach to solve the Gaussian prior problem. This method combined variational graph auto-encoder with diffusion models to advance the learning of more pliable prior distributions.

2.2. Link Prediction Methods for Multiplex Networks

In terms of link prediction in multiplex networks, Gallo et al. [32] proposed MultiSAGE, a generalization of GraphSAGE [33] that enables embeddings in multiplex networks and effectively integrates transitions between intra-layer and inter-layer edges. Nasiri et al. [34] proposed a local random walk based on an extended version of the pure random walk, called multiplex local random walk, to address the link prediction problem in multiplex networks. Bai et al. [35] regarded link prediction in multiplex networks as a multi-attribute decision-making problem, where potential links in the target layer were considered alternatives and the layers as attributes. The similarity of potential links in each layer was treated as an attribute value. The alternatives were ranked using the TOPSIS method, and the attributes were weighted by inter-layer correlation, achieving good prediction performance. Najari et al. [8] introduced a link prediction method based on inter-layer similarity for multiplex networks and proposed an inter-layer similarity metric that fused both inter-layer and intra-layer node features, thereby improving prediction accuracy. Shan et al. [36] proposed a supervised link prediction method in multiplex networks, which formulated link prediction as a binary classification problem. LISCNE [10] exploited both common and local features of multiplex networks while leveraging inter-layer similarity. Gao et al. [9] proposed a new embedding method called ILAR to capture and fully utilize multiplex networks with multiple relationships and complex layer correlations. ILAR employed two convolutional modules to capture sufficient complementarities and correlations in multiplex networks. Wang et al. [37] introduced the LPGRI method, which enhanced link prediction in multiplex networks by integrating intra-layer and inter-layer information, where the contribution of inter-layer information was quantified through a global relevance index.
Despite the remarkable effectiveness of previous link prediction methods for multiplex networks, their shortcomings still deserve our attention. In general, they do not sufficiently exploit the inter-layer relationships. As a result of the above elucidation, we are motivated to adopt the following method to better understand and explore the topology and potential links in multiplex networks. In comparison with existing models, the proposed LATGCN is capable of comprehensively exploiting both intra-layer and inter-layer information within multiplex networks.

3. Methodology

In this section, we first describe link prediction problem for multiplex networks and then provide a detailed overview of our model architecture, including explanations of its individual modules.

3.1. Problem Description of Link Prediction for Multiplex Networks

A multiplex network which consists of N nodes and k single-layer networks is defined as M G = G 1 , G 2 , , G k , where 0 < k K , and K denotes the total number of layers in the multiplex network. G k = V , ξ k denotes the kth layer network, where V = v 1 , v 2 , v N and E = ξ 1 , ξ 2 , , ξ k represent the set of N nodes and the set of edges in all layers respectively. The topology of this layer can be represented by an adjacency matrix A k . The embedding representation of node v i V in each relation is defined as h i R K × d 1 d 1 N . The global embedding representation of the node in all relations is defined as g i R d 2 d 2 N . The edge probability function is defined as p v i k , v j k 0 , 1 . The LATGCN model we propose is specifically designed to predict intra-layer links within a multiplex network. In particular, it leverages information from other layers to predict the links within the target layer G k .

3.2. The Proposed Model

The proposed LATGCN aims to capture the features of target nodes of a multiplex network carrying information from each layer. The architecture of the LATGCN model is shown in Figure 1, which consists of three major components, i.e., embedding representation extraction module, layer attention mechanism module, and embedding fusion module. Firstly, the embedding representation extraction module is designed to embed the topology information of the multiplex network into the node features. Secondly, the generated feature representations are fed into the layer attention mechanism module to obtain the layer importance scores. Thirdly, the extracted features and layer importance scores are processed by the embedding fusion module to obtain the final embedding representations. Fourthly, the decoder is used to generate the edge vectors and then the activation function is used to get the probabilities of the existence of edges. Finally, a loss function is used to supervise the entire training process of the model. In the following, we provide a detailed description of the LATGCN model.

3.2.1. Embedding Representation Extraction Module

The embedding representation extraction module aims to efficiently extract the node features of the relevant nodes of a multiplex network. As shown in Figure 2, LATGCN uses the output of Node2vec as the input to a two-layer GCN and obtains the embedding representations of the relevant nodes.
A multiplex network is defined as M G = G 1 , G 2 , , G k , where G k = V , ξ k denotes the kth layer ( 0 < k K ) . We represent the topology of each layer using an adjacency matrix A k , and use the adjacency matrices along with the initial features of the multiplex network as inputs to the GCN. Then, the task of embedding representation extraction for the kth-layer network is defined as:
H k l + 1 = σ ( D ˜ k 1 2 A ˜ k D ˜ k 1 2 H k ( l ) W k ( l ) ) ,
where A ˜ k = A k + I N k is the self-looped adjacency matrix of the kth-layer network G k . I N denotes the identity matrix, and D ˜ i i k = j A ˜ i j k denotes the diagonal matrix. W k l represents the trainable weight matrix at layer l for the kth-layer network. σ · represents the activation function which uses ReLU · in this paper. The input feature matrix at layer l is denoted as H k ( l ) = R N × d 0 , where N is the number of nodes and d 0 is the dimension of the initial embeddings. For the initial input H k ( 0 ) , we use the output of Node2vec.
Next, the overall processing equation of the multiplex network for the embedding representation extraction module is defined as:
H 1 l + 1 , H 2 l + 1 , , H k l + 1 = f A 1 , H 1 l ; A 2 , H 2 l ; ; A k , H k l ,
where H k l + 1 ( 0 < k K ) is the output of the kth-layer network after applying GCN. H k l represents the set of node embedding vectors learned by Node2vec on the kth-layer. A k is the adjacency matrix of the kth-layer multiplex network.
Inspired by GAT [38], this paper employs the layer attention mechanism to evaluate the contribution of nodes from other layers to the nodes in the target layer, after obtaining the node embedding representations by using the GCN model. This is achieved by analyzing the relationships among their neighbors across layers in the network. The mechanism assigns importance scores to the nodes in each layer, which indicates the importance of the nodes in other layers with respect to the nodes in the target layer, and accordingly adjusts the embedding matrix of each layer to ensure that information from important layers has a more significant impact on the global embedding representation.

3.2.2. Layer Attention Mechanism Module

In this paper, to capture the differences in information across layers relative to the target layer, LATGCN considers the influence of neighboring nodes on the target node. Specifically, it aggregates the effects of first-order neighbors, normalizes the per-layer influence, and assigns higher weights to layers where the neighboring nodes have greater impact. The schematic of the layer attention mechanism is shown in Figure 3.
The input to the layer attention mechanism is defined as h k = h 1 k , h 2 k , , h N k , where h i k R F is a set of node features and N is the number of nodes. F is the number of each node’s features. To obtain sufficient expressive power, we need to transform the input features into higher level features, which requires at least one learnable linear transformation. For this purpose, a shared linear transformation W k R F × F parameterized by a trainable weight matrix is applied to each node to perform self-attention., i.e., a kind of shared attention mechanism a : R F × R F R . Thus, the importance of the feature of node v j to node v i is defined as:
e i j k = a W k h i k , W k h j k ,
where e i j k is the attention coefficient of node v j to node v i .
In this paper, the layer attention mechanism is a single-layer feed-forward neural network parameterized by the weight vector a k R 2 F , and activated by LeakyReLU. After full expansion, the importance of the feature of node v j to node v i can be expressed as:
e i j k = LeakyReLU ( a k ) W k h i k h j k ,
where ⊤ denotes the transposition and ‖ denotes the concatenation operation.
In order to make the layer importance scores more easily comparable across layers, we normalize the choice of node v j for each layer of the multiplex network. The layer importance score is defined as:
α i k = j N i exp e i j k p = 1 P m N i exp e i m p ,
where α i k is the layer importance score of node v i at the kth-layer 0 < k K , 0 < α i k < 1 and N i is the first order neighbor of the node v i .
After calculating the layer importance score for each layer of node v i , LATGCN weights and fuses the layer importance scores with the node embedding representations, i.e., local embedding representations of the corresponding layers to obtain the global embedding representation g i .
In order to more accurately capture the complex relationship in the multiplex network, the global embedding is defined as:
g i = α i 1 h i 1 + α i 2 h i 2 + + α i K h i K ,
where g i is the global embedding representation of node v i . h i = h i 1 , h i 2 , , h i K ( h i R F ) is the embedding representation of node v i in each layer and K is the number of layers of the multiplex network.

3.2.3. Embedding Representation Fusion Module

In multiplex networks, node relationships within the target layer and node relationships in other layers are considered. The local embedding representation focuses on the node relationships and structural features of the target layer, while the global embedding representation integrates the information in the multiplex network. Therefore, the final embedding representation must be computed based on the local embedding representation and the global embedding representation. Then, a embedding representation fusion module is proposed, which introduces a learnable parameter to weight the local and the global embedding representation, thereby producing the final node embedding in the target layer. The parameter β is automatically adjusted during the training process to optimally balance the contributions of local information and global information. The final embedding representation is defined as:
H i k = β h i k + 1 β g i ,
where h i k denotes the local embedding representation of node v i at the target layer k 0 < k K . H i k is the final embedding representation of the multiplex network at the target layer k for node v i . The parameter β ( 0 < β < 1 ) is trainable and is initialized using a uniform distribution.
Finally, the obtained final embedding representations of the target layer nodes v i and v j are used as inputs to the decoder. We perform a Hadamard accumulation operation on the final embedding representations H i k and H j k to obtain the edge vector H i j k , as shown in Equation (8). The sigmoid activation function is then used to obtain the probability p that the link exists, as shown in Equation (9).
H i j k = H i k H j k .
p = Sigmoid H i j k = 1 1 + exp H i j k .

3.2.4. Model Optimization

Our goal is to predict the adjacency matrix of the target layer k based on the existing adjacency matrix of the multiplex network. We set the edge labels present in the target layer to 1 and the edge labels absent to 0. Supervised learning is carried out and the BCEWithLogitLoss is used to measure the difference between the true value and the predicted value. The loss function and the activation function are respectively shown in Equations (10) and (11).
L ( s , y ) = y log σ ( s ) + ( 1 y ) log 1 σ ( s ) ,
σ ( s ) = 1 1 + e s ,
where s is the logit score predicted by the model and y is the true label (0 or 1). σ ( · ) is the activation function that converts the logit to a probability value.

3.3. Complexity Analysis

The inputs to the proposed model LATGCN are the multiplex network, the initial features, the maximum number of iterations, and the patience for early stopping. The model consists of two layers of GCN, a layer attention mechanism, and an embedding fusion module. In the final stage, a decoder followed by an activation function is applied to estimate the probability of link existence. Its output is the predicted adjacency matrix of the target layer.
For ease of analysis, we clarify the meaning of the symbols here. We define N, E m a x , F 0 , F 1 , F 2 , F 3 , K, M, and J as the number of nodes of the multiplex network, the maximum value of the number of edges in each layer of the multiplex network, the initial feature dimensionality, the first layer GCN output dimensionality, the second layer GCN output dimensionality, the layer attention output dimensionality, the number of layers of the multiplex network, the number of predicted edges, and the number of epochs for training, respectively. To compute the complexity of LATGCN, we focus on the local embedding representation extraction module, the layer attention mechanism module, the embedding representation fusion module, the decoder part, and the loss function part.
After a comprehensive analysis of our proposed LATGCN, the local embedding representation extraction module uses two-layer GCN and the time complexity of it is O K E m a x F 0 + N F 1 + E m a x F 1 + N F 2 . The time complexity of the layer attention mechanism module is O K E m a x F 2 + E m a x F 3 . The time complexity of the embedding representation fusion module is O K N F 2 . The time complexity of the decoder part is O M F 2 . The time complexity of the loss function part is O ( M ) . As F 0 , F 1 , F 2 , and F 3 increase, both runtime and memory requirements grow. When F 0 and F 1 are relatively large, the two-layer GCN has a greater impact on runtime, whereas when F 2 and F 3 is large, the layer attention mechanism exerts a stronger influence on runtime. Since F 0 , F 1 , F 2 and F 3 are closer to each other, the overall time complexity can be simplified as O J K E m a x F + N F + M F , where F = max F 1 , F 2 , F 3 . Obviously, M < E m a x . For the sparser layer, M < E m a x N ; for the denser layer, E m a x > N . Thus, for sparser layers, the time complexity per iteration is O ( K F N ) , while for denser layers, it is O ( K E m a x F ) . As the number of layers K in a multiplex network increases, the time complexity grows, thereby increasing computational costs.

4. Experiment

In this section, we conduct a series of experiments to evaluate the performance of our proposed LATGCN model. Firstly, the datasets, the experimental settings, the evaluation metrics, and the baseline methods are described. Secondly, a parameter sensitivity analysis of LATGCN will be performed. Finally, the proposed LATGCN is compared with several state-of-the-art methods and the experimental results are analyzed.

4.1. Datasets

In this paper, we conducted experiments on six widely used public network datasets, which include CS-Aarhus [39], CKM-Physicians-Innovation [40], Twitter-Foursquar [41,42], ff-tw-yt [43], Rattus [44], and Sacchpomb [44]. These datasets are described below and the statistics of these datasets are shown in Table 1.
CS-Aarhus (CS): This multiplex network consists of 61 nodes and five types of edges. The nodes are staff members of the Department of Computer Science at Aarhus University. These types of edges are Facebook, Casual, Work, Collaboration, and Lunch.
CKM-Physicians-Innovation (CKM): This multiplex network consists of 246 nodes and three types of edges. The nodes are physicians from four different towns in the United States, and the edges capture their interactions related to the adoption of new drugs. These types of edges are seeking advice, discussing cases, and building friendships.
Twitter-Foursquar (TF): This multiplex network consists of 1564 nodes and two types of edges, corresponding to user interactions on Twitter and Foursquare.
ff-tw-yt (FF): This multiplex network consists of 6407 nodes and three types of edges, representing user interactions on FriendFeed, Twitter, and YouTube.
Rattus (RA): This multiplex network consists of 2640 nodes and six types of edges, indicating the genetic interactions in Rattus Norvegicus. These types of edges are Physical association, Direct interaction, Colocalization, Association, Additive genetic interaction defined by inequality, and Suppressive genetic interaction defined by inequality. Here we only use its first three layers.
Sacchpomb (SA): This multiplex network consists of 4092 nodes and seven types of edges, indicating the genetic interactions in Saccharomyces Pombe. These types of edges are Direct interaction, Colocalization, Physical association, Suppressive genetic interaction defined by inequality, Synthetic genetic interaction defined by inequality, Additive genetic interaction defined by inequality, and Association.

4.2. Experimental Settings

For each dataset, we selected one layer as the target layer and used the remaining layers as auxiliary layers. For the target layer, 10% of edges were randomly selected as the test set, another 10% as the validation set, and the remaining edges were used for training. We employed early stopping to prevent overfitting and to save training time and computational resources. Layers with extremely low density are unlikely to contain meaningful patterns for learning, so we restricted the evaluation to layers with adequate connectivity. Specifically, for the RA dataset, due to the extreme sparsity of other layers, we only used the first three layers (i.e., Physical association, Direct interaction, and Colocalization) for link prediction; for the other datasets, link prediction was performed on all layers. All experiments were conducted using an NVIDIA RTX 4060 GPU, with Python 3.8.13 and PyTorch 2.1.2.

4.3. Evaluation Metrics

We use four widely used metrics to evaluate performance, including area under the curve (AUC) [45], F1 score [46], AP [47], and Accuracy [48]. For each metric, a higher value indicates better performance. A brief description of these metrics is provided below.
AUC: It is the area under the ROC curve, which is an important measure of the model’s ability to classify. It is essentially defined as the probability that the model assigns a higher probability score to a randomly selected positive instance than to a randomly selected negative one.
F1: It is a metric used to comprehensively evaluate a model’s precision and recall in classification tasks especially in scenarios with class imbalance. The equation of F1 metric is shown as:
F 1 = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l .
AP: It evaluates model performance by computing the area under the precision–recall curve, summarizing the trade-off between precision and recall at different thresholds.
Accuracy: It measures the proportion of correctly predicted links over all candidate links, providing an overall assessment of prediction correctness in link prediction tasks.

4.4. Baseline Methods

In order to conduct more comprehensive experiments, we compare the proposed LATGCN with classical link prediction methods and state-of-the-art link prediction methods.
NMF [49]: It learns potential relationships between nodes by decomposing the adjacency matrix into two low-rank non-negative matrices.
MGCN [50]: It is a multi-dimensional graph convolutional network framework that captures interactions both within and across layers.
LATGCN-VAR (VAR): It is a variant of our LATGCN, in which the layer importance score is uniformly set to 1 K , where K is the total number of layers in the multiplex network.
MNERLP-MUL (MNER) [51]: It is a state-of-the-art link prediction method for multiplex networks. It employs an aggregation model to encode information from different layers into a summarized weighted static network. This process considers the relative densities of each layer and integrates both local and global structural information, based on the correlations among merged nodes and edges.
HOPLP-MUL (HOP) [52]: It is a state-of-the-art link prediction method for multiplex networks. It aggregates information from multiple layers into a single weighted static network by considering the relative density of each layer. The method then iteratively estimates link probabilities by exploring longer paths between nodes, thereby capturing higher-order structural information.
Layer-aware Graph neural network (LAGNN) [53]: It is a state-of-the-art link prediction method for multiplex networks, which aggregates the adjacency matrices of different layers and measures the inter-layer similarity.

4.5. Parameter Sensitivity Analysis

Since the size of the layer attention dimension affects the quality of aggregated neighborhood information, and the quality of neighborhood information further influences the richness of feature representation captured by the model, we investigate the performance of the proposed LATGCN under different layer attention dimensions in link prediction tasks to identify the optimal dimension size. To ensure generalization, we conduct experiments on six real-world multiplex network datasets, where 10% of the edges in the target layer are randomly selected as the test set, another 10% to form the validation set, and the remaining edges are used for training. For overall performance evaluation, we compute the average metrics across all layers under different attention dimensions for each dataset.
The experimental results for the metrics under different layer attention dimensions are presented in Figure 4. For the AUC metric, the CKM dataset shows an increasing trend as the layer attention dimension increases, reaching its optimal value at 256. In contrast, for the CS, TF, FF, RA, and SA datasets, AUC achieves the best performance at 128, but decreases when the dimension reaches 256, with the most pronounced drop observed in the CS dataset. For the F1 metric, the SA dataset exhibits an increasing trend with larger attention dimensions. For the CS, CKM, TF, FF, and RA datasets, F1 achieves the best performance at 128, the worst at 64, and then declines again at 256, with the CS dataset showing the most significant decrease. For the AP metric, the CKM and SA datasets show improvements with larger attention dimensions, performing worst at 32. For the CS, TF, FF, and RA datasets, AP achieves its highest value at 128 but drops at 256, with the CS dataset again showing a clear decline. For the Accuracy metric, the SA dataset shows improvements with increasing dimensions, while the CS, CKM, FF, and RA datasets achieve the best performance at 128. At 256, Accuracy decreases significantly on the CS dataset, whereas the TF dataset remains unchanged.
From the above analysis, we observe that the SA dataset benefits from larger attention dimensions, as its relatively large scale allows the model to capture richer neighborhood information. Conversely, the CS dataset suffers a significant performance drop at 256 due to its smaller scale, where larger attention dimensions may introduce noise, thereby degrading link prediction performance. Overall, our model achieves the best performance when the layer attention dimension is set to 128. Therefore, in the subsequent experiments, we fix the attention dimension to 128 to ensure optimal performance.

4.6. Comparison of the Link Prediction Performance

To comprehensively compare the performance of different methods, we compare the time complexity of the NMF, MGCN, MNER, HOP, LAGNN, and our LATGCN methods. As shown in Table 2, where r is the factorization dimension of NMF, m 1 and m 2 are the hidden layer dimension and the projection matrix dimension in the MGCN model, respectively. d 0 , d 1 , and d 2 are the input layer, hidden layer, and output layer dimensions of the two-layer GCN in LAGNN, respectively. D a v g is the average of the node degree; N is the number of nodes in MNER and HOP. E m a x is the maximum value of the number of edges in each layer of the multiplex network, and F represents the maximum dimensionality of the feature vectors in LATGCN. Obviously, NMF, MGCN, MNER, HOP and LAGNN exhibit significantly higher time complexity than LATGCN on larger datasets with more nodes, particularly when the number of nodes is larger than the maximum number of edges per layer. Moreover, under identical conditions, MNER demonstrates a higher time complexity compared to HOP.
To better analyze how the time complexity of the proposed LATGCN changes with dataset scale, we compare its actual runtime across datasets of different scale, as shown in Figure 5. In this experiment, the number of training epochs is fixed at 500, the layer attention dimension is set to 128. We compare the overall runtime of LATGCN on multiplex networks of varying scale. As shown in Figure 5, the horizontal axis represents the dataset scale. The dataset scale is measured using log ( K N ) , where K is the number of layers in multiplex network, N is the number of nodes in multiplex network. The vertical axis represents runtime, measured in minutes. The results show that the runtime of LATGCN increases with dataset scale, particularly when both the number of nodes and the number of layers are large.
Next, the performance of the above methods is compared in the link prediction task. For each dataset, 10% of the edges in the target layer are randomly removed to form the test set, another 10% to form the validation set, and the remaining edges are used for training. Link prediction is conducted independently on each layer of the multiplex networks, and the corresponding performance results are presented in the following section.
As shown in Table 3, the AUC values achieved by our method are generally higher than those of the baseline methods across all layers, except for the first layer of the CS dataset and the entire TF dataset. Similarly, Table 4 shows that our method outperforms the baseline methods in terms of F1 scores on all layers, except for the first layer of the CS dataset and the sixth layer of the SA dataset.
As for the AP metric, Table 5 shows that our proposed LATGCN method achieves superior results across all datasets, particularly on the TF dataset. As for the Accuracy metric, Table 6 shows that our proposed LATGCN method achieves superior results across all datasets, especially for the third layer of the CS dataset.
Although the AUC metrics of MNER and HOP are higher than that of our method in the TF dataset, the time complexity of them is higher for larger datasets such as the TF dataset. Although our method LATGCN performs slightly worse than MNER and HOP on the TF dataset, it requires less running time on larger datasets, thereby reducing the overall computational cost. Furthermore, as shown in Table 4, Table 5, Table 6, LATGCN achieves much higher F1, AP, and Accuracy metrics on the TF dataset, indicating that it can better balance precision and recall, especially in the larger datasets. The higher accuracy reflects the LATGCN’s capability of making more precise and consistent link predictions. More notably, as shown in Table 5, LATGCN achieves significantly higher AP metrics than VAR on the TF dataset, indicating the superior effectiveness of our proposed layer attention mechanism. This improvement is mainly attributed to the fact that when the number of layers in a multiplex network is small, our layer attention mechanism assigns relatively larger importance scores to each layer, thereby enabling more effective aggregation of information across layers. On the TF dataset, the number of layers in the multiplex network is two, so LATGCN performs better when the number of layers is less.

5. Conclusions

In this paper, we propose a new deep learning algorithm, LATGCN, to address the multiplex network link prediction problem. First, an embedding representation extraction module is employed to capture node features. Second, a layer attention mechanism module synthesizes the information from each layer of a multiplex network using a layer importance score, which effectively captures the contribution of nodes in other layers to those in the target layer. Based on the learned contributions, the model assigns optimal weights to the embedding representations of each layer, enabling a more effective aggregation of global topological information in the multiplex network. Finally, we use an embedding representation fusion module to merge the local embedding representation and the global embedding representation. We conducted extensive experiments on six public multiplex network datasets, and the results validate the superiority of LATGCN. In addition, we demonstrate that failing to effectively synthesize embedding representations across layers degrades model performance, whereas LATGCN overcomes this limitation and achieves strong results on real multiplex network datasets. Although LATGCN shows strong performance in multiplex network link prediction, it still has several limitations. In particular, the model may incur high computational and memory costs when applied to very large-scale multiplex networks.
In future work, we will extend our framework to link prediction tasks for multiplex signed networks, which will be a challenging research direction for link prediction in multiplex networks.

Author Contributions

Conceptualization, M.Y.; methodology, M.Y.; funding acquisition, M.Y.; resources, M.Y.; supervision, M.Y.; formal analysis, Y.H.; validation, Y.H.; software, Y.H.; visualization, Y.H.; investigation, Y.H.; writing—original draft preparation, M.Y. and Y.H.; writing—review and editing, M.Y. and Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant No. 62402327, Natural Science Foundation Joint Fund of Liaoning Province under Grant No. 2023-BSBA-244, and Liaoning Provincial Department of Education funding for research projects under Grant No. LJ212410142053.

Data Availability Statement

This study analyzed existing data that are publicly available. These data were derived from the following resources available in the public domain: CS-Aarhus, CKM-Physicians-Innovation, Rattus, and Sacchpomb (https://manliodedomenico.com/data.php, accessed on 7 July 2024), Twitter-Foursquare (https://github.com/shivansh-mishra/linkpredict-multiplex-layer, accessed on 5 July 2024), and ff-tw-yt (https://uuinfolab.github.io/data, accessed on 10 July 2024).

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Chen, D.; Zhang, S.; Zhao, Y.; Xie, M.; Wang, D. An Attribute Graph Embedding Algorithm for Sensing Topological and Attribute Influence. Mathematics 2024, 12, 3644. [Google Scholar] [CrossRef]
  2. He, Q.; Zhang, S.; Cai, Y.; Yuan, W.; Ma, L.; Yu, K. A Survey on Exploring Real and Virtual Social Network Rumors: State-of-the-Art and Research Challenges. ACM Comput. Surv. 2025, 57, 1–37. [Google Scholar] [CrossRef]
  3. He, Q.; Zhang, L.; Fang, H.; Wang, X.; Ma, L.; Yu, K.; Zhang, J. Multistage competitive opinion maximization with Q-learning-based method in social networks. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 7158–7168. [Google Scholar] [CrossRef]
  4. Lakshmi, T.J.; Bhavani, S.D. Link prediction approach to recommender systems. Computing 2024, 106, 2157–2183. [Google Scholar] [CrossRef]
  5. Abduljabbar, D.A.; Hashim, S.Z.M.; Sallehuddin, R. An enhanced evolutionary algorithm for detecting complexes in protein interaction networks with heuristic biological operator. In Proceedings of the Recent Advances on Soft Computing and Data Mining: Proceedings of the Fourth International Conference on Soft Computing and Data Mining (SCDM 2020), Melaka, Malaysia, 22–23 January 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 334–345. [Google Scholar]
  6. Wang, H.; Zhang, R.; Cheng, X.; Yang, L. Hierarchical traffic flow prediction based on spatial-temporal graph convolutional network. IEEE Trans. Intell. Transp. Syst. 2022, 23, 16137–16147. [Google Scholar] [CrossRef]
  7. Dong, S.; Wang, P.; Abbas, K. A survey on deep learning and its applications. Comput. Sci. Rev. 2021, 40, 100379. [Google Scholar] [CrossRef]
  8. Najari, S.; Salehi, M.; Ranjbar, V.; Jalili, M. Link prediction in multiplex networks based on interlayer similarity. Phys. A Stat. Mech. Its Appl. 2019, 536, 120978. [Google Scholar] [CrossRef]
  9. Gao, M.; Jiao, P.; Lu, R.; Wu, H.; Wang, Y.; Zhao, Z. Inductive link prediction via interactive learning across relations in multiplex networks. IEEE Trans. Comput. Soc. Syst. 2022, 11, 3118–3130. [Google Scholar] [CrossRef]
  10. Lu, R.; Jiao, P.; Wang, Y.; Wu, H.; Chen, X. Layer information similarity concerned network embedding. Complexity 2021, 2021, 2260488. [Google Scholar] [CrossRef]
  11. Newman, M.E. Clustering and preferential attachment in growing networks. Phys. Rev. E 2001, 64, 025102. [Google Scholar] [CrossRef] [PubMed]
  12. Dierk, S. The SMART retrieval system: Experiments in automatic document processing—Gerard Salton, Ed. (Englewood Cliffs, NJ: Prentice-Hall, 1971, 556 pp., $15.00). IEEE Trans. Prof. Commun. 1972, 1, 17. [Google Scholar] [CrossRef]
  13. Sorensen, T. A method of establishing groups of equal amplitude in plant sociology based on similarity of species content and its application to analyses of the vegetation on Danish commons. Biol. Skr. 1948, 5, 1–34. [Google Scholar]
  14. Ravasz, E.; Somera, A.L.; Mongru, D.A.; Oltvai, Z.N.; Barabási, A.L. Hierarchical organization of modularity in metabolic networks. Science 2002, 297, 1551–1555. [Google Scholar] [CrossRef]
  15. Adamic, L.A.; Adar, E. Friends and neighbors on the web. Soc. Netw. 2003, 25, 211–230. [Google Scholar] [CrossRef]
  16. Zhou, T.; Lü, L.; Zhang, Y.C. Predicting missing links via local information. Eur. Phys. J. B 2009, 71, 623–630. [Google Scholar] [CrossRef]
  17. Lü, L.; Jin, C.H.; Zhou, T. Similarity index based on local paths for link prediction of complex networks. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 2009, 80, 046122. [Google Scholar] [CrossRef] [PubMed]
  18. Aziz, F.; Gul, H.; Muhammad, I.; Uddin, I. Link prediction using node information on local paths. Phys. A Stat. Mech. Its Appl. 2020, 557, 124980. [Google Scholar] [CrossRef]
  19. Katz, L. A new status index derived from sociometric analysis. Psychometrika 1953, 18, 39–43. [Google Scholar] [CrossRef]
  20. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  21. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar] [CrossRef]
  22. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
  23. Cao, S.; Lu, W.; Xu, Q. Grarep: Learning graph representations with global structural information. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management, Melbourne, Australia, 19–23 October 2015; pp. 891–900. [Google Scholar]
  24. Wang, D.; Cui, P.; Zhu, W. Structural deep network embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1225–1234. [Google Scholar]
  25. Xuan, Q.; Fang, B.; Liu, Y.; Wang, J.; Zhang, J.; Zheng, Y.; Bao, G. Automatic pearl classification machine based on a multistream convolutional neural network. IEEE Trans. Ind. Electron. 2017, 65, 6538–6547. [Google Scholar] [CrossRef]
  26. Niepert, M.; Ahmed, M.; Kutzkov, K. Learning convolutional neural networks for graphs. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 2014–2023. [Google Scholar]
  27. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  28. Wang, H.; Wang, J.; Wang, J.; Zhao, M.; Zhang, W.; Zhang, F.; Xie, X.; Guo, M. Graphgan: Graph representation learning with generative adversarial nets. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  29. Li, Z.; Zhang, L.; Song, G. Sepne: Bringing separability to network embedding. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 4261–4268. [Google Scholar]
  30. Yao, Y.; Guo, P.; Mao, Z.; Ti, Z.; He, Y.; Nian, F.; Zhang, R.; Ma, N. Multi-scale contrastive learning via aggregated subgraph for link prediction. Appl. Intell. 2025, 55, 489. [Google Scholar] [CrossRef]
  31. Su, H.; Li, Z.; Yuan, C.A.; Vladimir, F.F.; Huang, D.S. Variational graph neural network with diffusion prior for link prediction. Appl. Intell. 2025, 55, 90. [Google Scholar] [CrossRef]
  32. Gallo, L.; Latora, V.; Pulvirenti, A. MultiSAGE: A multiplex embedding algorithm for inter-layer link prediction. arXiv 2022, arXiv:2206.13223. [Google Scholar]
  33. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  34. Nasiri, E.; Berahmand, K.; Li, Y. A new link prediction in multiplex networks using topologically biased random walks. Chaos Solitons Fractals 2021, 151, 111230. [Google Scholar] [CrossRef]
  35. Bai, S.; Zhang, Y.; Li, L.; Shan, N.; Chen, X. Effective link prediction in multiplex networks: A TOPSIS method. Expert Syst. Appl. 2021, 177, 114973. [Google Scholar] [CrossRef]
  36. Shan, N.; Li, L.; Zhang, Y.; Bai, S.; Chen, X. Supervised link prediction in multiplex networks. Knowl.-Based Syst. 2020, 203, 106168. [Google Scholar] [CrossRef]
  37. Wang, C.; Tang, F.; Zhao, X. LPGRI: A global relevance-based link prediction approach for multiplex networks. Mathematics 2023, 11, 3256. [Google Scholar] [CrossRef]
  38. Velickovic, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. Stat 2017, 1050, 10–48550. [Google Scholar]
  39. Magnani, M.; Micenkova, B.; Rossi, L. Combinatorial analysis of multiple networks. arXiv 2013, arXiv:1303.4986. [Google Scholar] [CrossRef]
  40. Coleman, J.; Katz, E.; Menzel, H. The diffusion of an innovation among physicians. Sociometry 1957, 20, 253–270. [Google Scholar] [CrossRef]
  41. Torabi, E.; Ghobaei-Arani, M.; Shahidinejad, A. Data replica placement approaches in fog computing: A review. Clust. Comput. 2022, 25, 3561–3589. [Google Scholar] [CrossRef]
  42. Yang, R.; Yang, C.; Peng, X.; Rezaeipanah, A. A novel similarity measure of link prediction in multi-layer social networks based on reliable paths. Concurr. Comput. Pract. Exp. 2022, 34, e6829. [Google Scholar] [CrossRef]
  43. Dickison, M.E.; Magnani, M.; Rossi, L. Multilayer Social Networks; Cambridge University Press: Cambridge, UK, 2016. [Google Scholar]
  44. De Domenico, M.; Nicosia, V.; Arenas, A.; Latora, V. Structural reducibility of multilayer networks. Nat. Commun. 2015, 6, 6864. [Google Scholar] [CrossRef]
  45. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef]
  46. Schütze, H.; Manning, C.D.; Raghavan, P. Introduction to Information Retrieval; Cambridge University Press Cambridge: Cambridge, UK, 2008; Volume 39. [Google Scholar]
  47. Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
  48. Liben-Nowell, D.; Kleinberg, J. The link prediction problem for social networks. In Proceedings of the Twelfth International Conference on Information and Knowledge Management, New Orleans, LA, USA, 2–8 November 2003; pp. 556–559. [Google Scholar]
  49. Lin, C.J. Projected gradient methods for nonnegative matrix factorization. Neural Comput. 2007, 19, 2756–2779. [Google Scholar] [CrossRef]
  50. Ma, Y.; Wang, S.; Aggarwal, C.C.; Yin, D.; Tang, J. Multi-dimensional graph convolutional networks. In Proceedings of the 2019 Siam International Conference on Data Mining, SIAM, Calgary, AB, Canada, 2–4 May 2019; pp. 657–665. [Google Scholar]
  51. Mishra, S.; Singh, S.S.; Kumar, A.; Biswas, B. MNERLP-MUL: Merged node and edge relevance based link prediction in multiplex networks. J. Comput. Sci. 2022, 60, 101606. [Google Scholar] [CrossRef]
  52. Mishra, S.; Singh, S.S.; Kumar, A.; Biswas, B. HOPLP- MUL: Link prediction in multiplex networks based on higher order paths and layer fusion. Appl. Intell. 2023, 53, 3415–3443. [Google Scholar] [CrossRef]
  53. Abedini, M.; Shakibian, H. Inter-layer similarity-based graph neural network for link prediction in social multiplex networks. In Proceedings of the 2024 10th International Conference on Web Research (ICWR), Tehran, Iran, 24–25 April 2024; IEEE: New York, NY, USA, 2024; pp. 76–80. [Google Scholar]
Figure 1. Overall framework of the proposed LATGCN model. (1) Input: a multiplex network; (2) embedding representation extraction module: learns node embeddings in each layer through two-layer GCN; (3) layer attention mechanism module: assigns higher weights to layers with greater impact on neighboring nodes; (4) embedding representation fusion module: fuses local and global embedding representations.
Figure 1. Overall framework of the proposed LATGCN model. (1) Input: a multiplex network; (2) embedding representation extraction module: learns node embeddings in each layer through two-layer GCN; (3) layer attention mechanism module: assigns higher weights to layers with greater impact on neighboring nodes; (4) embedding representation fusion module: fuses local and global embedding representations.
Mathematics 13 03803 g001
Figure 2. Embedding extraction process in the LATGCN model. (1) Node2vec is used to extract the initial feature matrix for the multiplex network; (2) the initial feature matrix is then processed by a two-layer GCN.
Figure 2. Embedding extraction process in the LATGCN model. (1) Node2vec is used to extract the initial feature matrix for the multiplex network; (2) the initial feature matrix is then processed by a two-layer GCN.
Mathematics 13 03803 g002
Figure 3. Layer attention mechanism in LATGCN. (1) It aggregates the effects of first-order neighbors; (2) it assigns higher weights to layers where the neighboring nodes have greater impact.
Figure 3. Layer attention mechanism in LATGCN. (1) It aggregates the effects of first-order neighbors; (2) it assigns higher weights to layers where the neighboring nodes have greater impact.
Mathematics 13 03803 g003
Figure 4. Comparison of metrics under different layer attention dimensions. (a) AUC; (b) F1; (c) AP; (d) Accuracy.
Figure 4. Comparison of metrics under different layer attention dimensions. (a) AUC; (b) F1; (c) AP; (d) Accuracy.
Mathematics 13 03803 g004
Figure 5. Overall runtime of LATGCN with varying dataset scale.
Figure 5. Overall runtime of LATGCN with varying dataset scale.
Mathematics 13 03803 g005
Table 1. Statistics of datasets.
Table 1. Statistics of datasets.
DatasetNodesEdgesLayers
CS616205
CKM24615513
TF156430,8822
FF640730,8823
RA264042293
SA409263,6777
Table 2. Comparison of time complexity.
Table 2. Comparison of time complexity.
Algorithm NameTime Complexity
NMF O ( K | N | 2 r )
MGCN O ( K 2 | m 1 | 2 m 2 )
MNER O ( | N | 2 D avg 2 )
HOP O ( | N | 2 D avg )
LAGNN O ( N d 1 ( d 0 + d 2 ) )
LATGCN O ( K N F ) s . t . M < E max N
O ( K E max F ) s . t . E max > N
Table 3. Comparison of AUC on each layer of the six datasets.
Table 3. Comparison of AUC on each layer of the six datasets.
DatasetLayerNMFMGCNMNERHOPLAGNNVARLATGCN
CS10.66340.85000.95000.95000.79000.84250.8825
20.81940.95270.92310.92310.83430.95270.9645
30.75001.00000.33330.33331.00001.00001.0000
40.64060.88890.77780.66670.83950.90120.9136
50.67170.82000.75000.75000.74500.82100.8350
CKM10.65830.82900.51110.48890.82010.83530.8405
20.71390.85430.48000.50000.81690.91690.9237
30.66950.86890.41860.41860.81310.86240.8758
TF10.36230.79631.00001.00000.84010.74010.8862
20.36950.82001.00001.00000.82000.74790.8717
FF10.68420.95160.25860.24140.75610.98620.9917
20.54080.83310.79170.60730.86470.88300.9014
30.69450.77560.80550.52740.81010.83250.8455
RA10.47120.92170.24100.20860.76090.93280.9537
20.63300.94810.02020.03030.72410.95790.9746
30.63641.00000.16670.33330.81661.00001.0000
SA10.78680.92010.30600.25370.82130.94090.9660
20.70270.97980.32430.24320.84240.98870.9994
30.69400.91500.64760.61600.70360.93840.9410
40.62270.76020.44240.41310.80080.86360.8893
50.92110.81600.62990.64170.88140.92310.9350
60.65510.80000.77050.72620.76370.84130.8606
70.80950.94270.36360.45450.91060.95130.9757
Table 4. Comparison of F1 on each layer of the six datasets.
Table 4. Comparison of F1 on each layer of the six datasets.
DatasetLayerNMFMGCNMNERHOPLAGNNVARLATGCN
CS10.73910.71700.87400.87200.71790.78430.8263
20.78260.74290.88670.89160.77420.78790.8966
30.66670.85710.66030.66260.75000.85711.0000
40.57140.72000.83230.78950.76190.93650.9474
50.75560.70370.74610.73720.76190.80110.8163
CKM10.73210.72580.72980.72120.70370.81740.8348
20.77050.73330.71380.72610.70100.87300.8837
30.73080.75760.69250.69330.70070.82140.8430
TF10.26400.70380.50000.50000.71330.75570.7648
20.28590.72600.50000.50000.73770.75150.7705
FF10.50000.97520.62930.62070.63930.96120.9883
20.58500.81850.81150.79100.74900.81850.8254
30.73960.79450.78330.75740.73060.78700.7953
RA10.33490.72400.57760.59700.66520.91950.9406
20.07840.69880.50940.51440.66250.94550.9550
30.42860.66670.58330.66660.66671.00001.0000
SA10.67650.72280.65270.62650.76790.92390.9330
20.48980.82000.66210.62160.73910.95240.9880
30.33920.74800.78030.76440.62320.88820.8990
40.50160.68920.69390.67970.74010.81860.8358
50.86450.68750.81380.81980.81360.87240.8770
60.55350.72270.83500.83180.69970.78490.7884
70.50000.82140.68170.72720.84000.94120.9588
Table 5. Comparison of AP on each layer of the six datasets.
Table 5. Comparison of AP on each layer of the six datasets.
DatasetLayerNMFMGCNMNERHOPLAGNNVARLATGCN
CS10.60250.81840.05340.05240.81700.84740.8817
20.80780.94910.04410.04710.83610.95410.9682
30.75001.00000.04080.06251.00001.00001.0000
40.60520.81820.03370.03720.86490.80420.8394
50.60180.79410.03370.03160.68360.78330.8040
CKM10.59590.77360.01540.01620.78380.78450.7936
20.63800.84110.01580.01800.81320.90340.9139
30.60450.83470.01840.01930.83630.83980.8430
TF10.39610.75080.00120.00120.85490.65460.8874
20.39980.78740.00120.00120.83610.66340.8488
FF10.68420.97560.01000.01140.79920.98420.9909
20.58500.77270.00140.00370.88270.87180.8831
30.60810.69320.00280.00850.78720.78520.7990
RA10.46270.88670.00020.00110.77340.89180.9227
20.65440.92050.00040.00060.76440.95830.9587
30.63641.00000.01030.01860.86261.00001.0000
SA10.79100.92650.00740.00620.83480.90670.9322
20.70270.97970.01660.01400.86710.99010.9994
30.56830.89640.00060.00060.73200.89810.9096
40.51660.73880.00150.00140.81560.82040.8424
50.88210.83700.00840.00870.87620.87680.8876
60.53680.75430.00280.00430.78650.79240.8138
70.80950.94740.00490.01270.92360.95830.9741
Table 6. Comparison of Accuracy on each layer of the six datasets.
Table 6. Comparison of Accuracy on each layer of the six datasets.
DatasetLayerNMFMGCNMNERHOPLAGNNVARLATGCN
CS10.68420.62500.19010.12410.72500.72500.7898
20.79170.65380.14490.14590.73080.73080.8846
30.75000.83330.02580.02190.66670.83331.0000
40.62500.61110.07150.03340.72220.93220.9445
50.71050.60000.10280.07000.75000.75560.7750
CKM10.65910.64580.01320.01930.66670.78130.8021
20.71430.64910.01440.01050.74560.85960.8684
30.66670.68630.01220.01070.59800.78870.8138
TF10.35440.58510.00120.00120.63010.68990.7012
20.35800.63430.00120.00120.70680.68320.7114
FF10.66670.97580.00480.00810.64520.96080.9884
20.56270.78640.03420.02930.71590.78670.7927
30.69410.75030.02980.03580.71360.74270.7527
RA10.47650.62260.00290.00220.50000.91390.9371
20.52040.59320.00010.00020.50450.94550.9546
30.63640.50010.00260.04170.50001.00001.0000
SA10.75190.63010.00500.00290.76920.92010.9290
20.66220.78050.00760.00460.70730.95120.9879
30.51650.67190.01400.01730.65780.88020.8902
40.55500.59210.00240.00330.74600.80510.8122
50.87000.58290.02100.01700.80520.86670.8708
60.59820.63070.01160.02290.72410.73210.7369
70.66670.79160.04670.01060.83330.93750.9570
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, M.; He, Y. A Link Prediction Algorithm Based on Layer Attention Mechanism for Multiplex Networks. Mathematics 2025, 13, 3803. https://doi.org/10.3390/math13233803

AMA Style

Yang M, He Y. A Link Prediction Algorithm Based on Layer Attention Mechanism for Multiplex Networks. Mathematics. 2025; 13(23):3803. https://doi.org/10.3390/math13233803

Chicago/Turabian Style

Yang, Mingzhou, and Yongqi He. 2025. "A Link Prediction Algorithm Based on Layer Attention Mechanism for Multiplex Networks" Mathematics 13, no. 23: 3803. https://doi.org/10.3390/math13233803

APA Style

Yang, M., & He, Y. (2025). A Link Prediction Algorithm Based on Layer Attention Mechanism for Multiplex Networks. Mathematics, 13(23), 3803. https://doi.org/10.3390/math13233803

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop