Next Article in Journal
MTD-YOLO: An Improved YOLOv8-Based Rice Pest Detection Model
Previous Article in Journal
A Three-Layer Sequential Model Predictive Current Control for NNPC Four-Level Inverters with Low Common-Mode Voltage
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ANHNE: Adaptive Multi-Hop Neighborhood Information Fusion for Heterogeneous Network Embedding

College of Electronic Engineering, National University of Defense Technology, Hefei 230031, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(14), 2911; https://doi.org/10.3390/electronics14142911
Submission received: 8 June 2025 / Revised: 14 July 2025 / Accepted: 18 July 2025 / Published: 21 July 2025
(This article belongs to the Special Issue Advances in Learning on Graphs and Information Networks)

Abstract

Heterogeneous information network (HIN) embedding transforms multi-type nodes into low-dimensional vectors to preserve structural and semantic information for downstream tasks. However, it struggles with multiplex networks where nodes connect via diverse semantic paths (metapaths). Information fusion mainly improves the quality of node embedding by fully exploiting the structure and hidden information within the network. Current metapath-based methods ignore information from intermediate nodes along paths, depend on manually defined metapaths, and overlook implicit relationships between nodes sharing similar attributes. Our objective is to develop an adaptive framework that overcomes limitations in existing metapath-based embedding (incomplete information aggregation, manual path dependency, and ignorance of latent semantics) to learn more discriminative embeddings. We propose an adaptive multi-hop neighbor information fusion model for heterogeneous network embedding (ANHNE), which: (1) autonomously extracts composite metapaths (weighted combinations of relations) via a multipath aggregation matrix to mine hierarchical semantics of varying lengths for task-specific scenarios; (2) projects heterogeneous nodes into a unified space and employs hierarchical attention to selectively fuse neighborhood features across metapath hierarchies; and (3) enhances semantics by identifying potential node correlations via cosine similarity to construct implicit connections, enriching network structure with latent information. Extensive experimental results on multiple datasets show that ANHNE achieves more precise embeddings than comparable baseline models.

1. Introduction

Network representation learning (NRL) [1] effectively embeds information networks into a compact and continuous low-dimensional space. It enables the extraction of valuable insights from networked systems while preserving their structural features, intrinsic properties, and semantic information. NRL supports a range of downstream semantic computing applications, including link prediction [2] and node classification [3], leading to applications in fields like recommender systems [4], social influence analysis [5], and beyond. Notably, DeepWalk [6], LINE [7], and node2vec [8] have been pioneering in integrating deep learning with NRL.
In recent years, the emergence of graph neural networks (GNNs) [9,10,11,12,13,14] has revolutionized the processing of non-Euclidean spatial data, facilitating its direct incorporation into neural networks. This development has facilitated the learning of embedded representations of complex networks enriched with feature information. GNNs employ either spectral methods [10,11] or methods based in the spatial domain [9,12] to compile data from the immediate neighbors of a specific node, enhancing the capabilities for structural analysis and local data integration within graph contexts. The introduction of the attention mechanism, as seen in the graph attention network (GAT) [15], enables selective learning of the importance of single-layer neighbors based on specific tasks, significantly enhancing performance across various downstream tasks. However, this single-layer approach is limited to aggregating information from only the one-hop neighbors.
In practice, numerous connections exist between different objects. Neglecting this heterogeneous information can obscure the distinct characteristics of various node types and the intricate semantic relationships among them, adversely affecting the efficacy of subsequent tasks. To tackle this challenge, multiplexed network representations [16,17,18,19,20,21] leverage metapaths to depict various composite links between nodes. Ref. [16] has extended the GAT model to amalgamate representations from different metapaths, allowing the model to assimilate more comprehensive semantic information. GATNE [22], a trailblazer, introduced AMHEN to NRL. Furthermore, GTN [23] exploits the Transformer architecture to autonomously learn and control the length of useful multi-hop metapaths based on edge-type relations. MHGCN [24] presents a single-layer attributed multi-order graph convolutional network that autonomously explores various metapaths for multi-hop information aggregation learning through homogeneous supervised guidance.
Although metapath-based HIN embedding methods have achieved state-of-the-art results in various tasks, they still suffer from several limitations:
(1)
Inadequate information mining along metapaths: Most metapath-based heterogeneous network representations [16,25,26] focus only on the information of the nodes at the two ends of the metapath and ignore the information of the intermediary nodes, which are also crucial for representation learning. As shown in Figure 1b, two authors in the ‘APA’ metapath can collaborate on different papers (A2 and A3), and neglecting the information related to these papers hampers the accurate mining of the relation between the authors. Some state-of-the-art approaches [18,27,28] address this by emphasizing message passing to capture various relations between nodes. However, these methods rely on stacking multiple layers of GNNs to aggregate neighbor information, which can lead to performance degradation and excessive smoothing.
(2)
The common method of determining walking paths through predefined metapaths introduces bias in network structure construction. On the one hand, the learning of network structure heavily depends on the quality of manually designed metapaths, which must be specified in advance based on experience or domain knowledge (e.g., the ‘APA’ and ‘PCP’ metapaths in Figure 1c, which indicate that two authors collaborated on a paper and two papers were presented at the same conference, respectively). Yet, they cannot autonomously extract valuable metapaths tailored to specific task scenarios. Especially for unknown HINs, valid metapaths cannot always be pre-identified [29]. Differently defined metapaths in the same dataset affect GCN performance due to varying semantics and values across different scenarios. On the other hand, the lack of effective integration of different metapaths is not conducive to extracting the network structure for learning. To illustrate, ‘APA’ signifies a scenario where two authors have co-authored a paper, whereas ‘APCPA’ indicates that articles by two authors are featured in the same conference or journal. Different metapaths represent diverse meanings and have varying importance, making their effective integration crucial for improvement.
(3)
Most metapath-based embedding models ignore potential semantic associations between node attributes. Current studies [17,18,30] only learn network embedding representations from primitive network structures. While the primitive structure reflects explicit node relations, it does not encompass all possible relations, missing many implicit relations between nodes. For instance, as depicted in Figure 1b, interdisciplinary papers P1 and P4 may not directly cite each other and lack connecting metapaths owing to their authors and topics originating from different fields. However, they share the same heterogeneous network embedding technique, indicating a latent semantic relation. Capturing these potential correlations between nodes allows for the exploration of deeper semantic relations, resulting in a more comprehensive network structure and improved node representation.
To systematically address limitations (1)–(3), we formulate three research problems (RPs) as the key objectives of this work:
(RP1). How do we adaptively construct multi-length metapaths based on heterogeneous relations? (Section 3.3 and Section 3.4)
(RP2). How do we selectively fuse multi-hop neighborhood information? (Section 3.5)
(RP3). How do we uncover latent semantics ignored by metapaths? (Section 3.6)
To address the aforementioned problems, we propose the adaptive multi-hop neighbor information fusion for heterogeneous network embedding (ANHNE) model. Specifically, we first introduce a composite metapath autonomous extraction (CMAE) model to autonomously learn combinations of sub-metapaths of various lengths without manual design. Based on the composite metapath, we propose a hierarchical semantic attention aggregation (HSAA) model, which selectively aggregates multi-hop structural semantic features to yield precise node representations on a hierarchical scale. Finally, a semantic information enhancement (SIE) model is designed to explore implicit correlations between nodes to unearth deep implicit associations based on the original network structure. An illustration of ANHNE is shown in Figure 2. The most noteworthy contributions of this paper are summarized as follows:
  • To the best of our knowledge, this paper presents the concept of composite metapaths for the first time. Unlike the commonly accepted notion of metapaths, composite metapaths are soft combinations of different relations for specific task scenarios, where the topological information of all nodes on the path can be aggregated through the matrix product of unipath relations with different weights.
  • We develop a hierarchical semantic attention aggregation model that autonomously determines the relative importance of valuable composite metapaths of specific lengths and selectively aggregates information across different hierarchical levels of neighboring nodes.
  • We propose a semantic information enhancement module that mines potential relations between nodes and extensively extracts implicit semantic relations without direct connectivity, enhancing feature representation learning to obtain high-quality representations.
  • We conduct extensive experiments on three real-world datasets to indicate the advantage of each module and demonstrate the superiority of our proposal to the state of the art for multiplex heterogeneous networks.
Figure 2. The overview framework of the proposed ANHNE.
Figure 2. The overview framework of the proposed ANHNE.
Electronics 14 02911 g002
The remainder of this paper is organized as follows: We discuss related works in Section 2 and briefly describe relevant preliminaries in Section 3. Section 3 also introduces our proposed method in detail. Then in Section 4, we present a series of experiments comparing our method with benchmark approaches and discuss the results. Finally, we conclude our work in Section 5.

2. Related Work

2.1. Graph Convolutional Networks (GCNs)

GCNs aim to derive low-dimensional node embeddings through the feature decomposition of graph Laplacian matrices [31]. Given a homogeneous graph G = { V , E } and the feature matrix X R V × d f of all nodes projected to the d f -dimension space, the feature transfer between layers can be formulated as
H ( l + 1 ) = σ ( D 1 / 2 ( A + I ) D 1 / 2 H ( l ) · W ( l ) )
where A R N × N is the adjacency matrix, where A i j = 1 if nodes i and j are connected. I denotes the identity matrix, and D is the degree matrix of A + I . W ( l ) is a trainable weight matrix. H ( l ) is the l-layer feature matrix, and H ( 0 ) = X . The depth of the GCN determines the extent to which the model can capture and transmit node features [32].
GraphSAGE [9], a large-scale GCN framework, employs an embedding function to sample and aggregate local neighborhood features of nodes. To selectively prioritize different neighborhood information, GAT introduced an attention-based message passing mechanism. AM-GCN [33] adaptively learns the depth correlation between topology and feature spaces using a multi-channel GCN approach. Scattering-GCN [34] integrates a geometric scattering transform with residual convolution to enhance traditional GCN models. Dual-light GCN [35] eliminates nonlinear embedding projections and aggregates information across multiple scales. Cross-GCN [36] introduces cross-feature map convolution to explicitly model cross-features of arbitrary orders. To streamline GCN models, SGC [37] removes the activation function, pre-computes the higher-order adjacency matrix, and uses it in a single GCN layer to approximate multi-layer information propagation. Zhang et al. [38] utilized the SGC paradigm to capture global cluster structures and adaptively select the appropriate convolution order for different graphs.
However, these initial GCN implementations were designed for homogeneous networks and failed to effectively preserve the heterogeneity and multiplicity of real-world network embeddings.

2.2. Heterogeneous Network Representation Learning

Efforts to enhance heterogeneous network representations using GNNs have been extensive. HGT [26] employs relation-dependent self-attention to model dynamic heterogeneous interactions, but ignores higher-order semantic connections beyond direct neighbors. HeCo [39] improves node embeddings by comparing metapath and network schema views, but requires predefined metapaths and misses implicit relations. DMGI [40] employs a consensus regularization scheme to address relations among specific node embedding types, incorporating perceptual nodes. PF-HIN [41] adopts a pre-training and fine-tuning framework for heterogeneous information network feature learning. MISE [42] uses mechanisms to maximize mutual information between local and global metapath representations, constructing graph structures through implicit node correlations.
The metapath approach is a classical method designed for heterogeneous graph models. HERec [43] employ randomized wandering sampling to construct walking paths on heterogeneous graphs for node embedding. PSHGNAN [44] integrates metapaths and meta-structures through an attention mechanism to learn node embeddings. However, the quality of empirically designed metapaths significantly impacts embedding performance. AMOGCN [45] fuses outputs from different depths of the graph convolutional layer to adaptively learn multi-length metapaths. Many metapath models only aggregate information from the endpoints, overlooking the relations between nodes during the message passing process. To effectively utilize information from intermediate nodes, Ref. [18] combines jointly supervised signals containing extrinsic and intrinsic mutual information, and [27] constructs a multi-behavioral graph convolutional network to represent multi-behavioral data. However, aggregating multi-hop neighbor information by stacking multiple GNN layers can lead to performance degradation and excessive smoothing. To counteract this, Xue et al. [46] aggregate neighbor information across different hops using an attentive mechanism, enabling selective extraction of important multi-hop features. MHNF [47] makes further improvements, allowing for the aggregation of information from multi-hop neighbor nodes in heterogeneous graphs. However, they only considered explicit relations in the network and lacked consideration of the natural homogeneity between node attributes.
To address the above problems, the ANHNE method introduced in this paper not only designs an adaptive learning strategy for multi-length metapaths based on the diverse relations among nodes during the information transfer process, but also considers the natural homogeneity among node attributes to effectively mine and aggregate multi-hop neighbor information in heterogeneous graphs. This capability represents a significant advantage of ANHNE over previous research efforts.

3. Methodology

3.1. Preliminaries

In this section, we summarize a concise overview of the essential definitions utilized throughout this paper in Table 1 for quick reference.
Definition 1.
Attributed multiplex heterogeneous network: It can be represented as G = G 1 , G 2 , , G r , , G | R | , where G r = { V , A r , X } is a unipath relation network with a particular relation r R , X R N × m denotes the attribute feature matrix, and A r { 0 , 1 } N × N is the adjacency matrix of the network G r . We denote the set of adjacency matrices as A = A 1 , A 2 , , A r , , A | R | . With a node type mapping function f v : V T and an edge type mapping function f e : E R , the node types and edge types in a heterogeneous network should satisfy T + R > 2 .
Definition 2.
Metapath: A metapath is usually formulated in the form of T 1 R 1 T 2 R 2 R l T l + 1 and abbreviated as T 1 T 2 T l + 1 . The path indicates that node 1 and l + 1 have a relation R = R 1 R 2 R l , where T i denotes the type of the node i, R j denotes the type of the edge j, and ∘ denotes the join operator in the relation.
Definition 3.
Composite metapath: Composite metapath is no longer an artificially specified path in Definition 2, but a soft combination of relations determined based on specific task scenarios. In this case, the combination relation R becomes R = R 1 R 2 R l = ( m 1 T w m 1 R m 1 )
( m 2 T w m 2 R m 2 ) ( m l T w m l R m l ) . Since the combination relation T 1 R 1 T 2 R 2 R l T l + 1 is able to learn different combinations of relations for a specific task, it can be used for mining richer and more valuable relations.

3.2. Overall Framework

In this section, we present the proposed ANHNE model for managing the attributed heterogeneous network, with a detailed depiction in Figure 2. Initially, we introduce the autonomous extraction module for composite metapaths, which autonomously identifies valuable combinations of sub-metapaths containing various relations and effectively integrates the information from each sub-metapath based on their assigned weights. Next, we propose a hierarchical semantic attention aggregation model to selectively aggregate semantic information from different hierarchical neighbors through composite metapaths. Finally, we construct a semantic information enhancement module to further explore implicit correlations between nodes and enrich the embedded network with more comprehensive semantic information. The proposed model is introduced and analyzed through RP1 (Section 3.3 and Section 3.4), RP2 (Section 3.5), and RP3 (Section 3.6):

3.3. Multipath Relation Aggregation

Denoted by { G r | r = 1 , 2 , , R } , each unipath relation network contributes differently in a given task scenario, and accurately determining their relative importance is crucial to reflecting the underlying relations accurately.
Considering scenarios such as multiple user–movie interactions on video streaming platforms, the decomposed unipath relation networks correspond to distinct types of relational links between users and movies. As illustrated in the U–M relation network in Figure 3, click and collect represent different edge types in the network, each exhibiting unique behavioral semantics. The varying importance of A 1 and A 2 reflects users’ differing behavioral preferences for movies at that time. Thus, multiple user–movie interactions with distinct relation semantics impact the network’s learning process, respectively. To capture these varied node dependencies based on unipath relation types, we propose a multipath relation aggregation approach. The multipath relation matrices of aggregating specific edge types with learned weights w m r are as follows:
A ˜ = r = 1 R w m r A r
where A r denotes the unipath relation adjacency matrix of relation R r , and the set of weights { w m r | r = 1 , 2 , R } is a set of trainable parameters that dynamically adjusts according to different tasks.

3.4. Composite Metapath Autonomous Extraction Model (CMAE)

Extracting neighborhood information through metapaths can significantly enhance the performance of embedding models. However, each manually specified metapath represents only a single type of relation, transferring interactions in a way that cannot capture the complex interactions among diverse node types. In practical scenarios, the relations between two nodes are often intricate and varied, with each relation playing distinct roles and possessing varying levels of importance across different tasks. Consequently, it is crucial to construct composite metapaths of specified lengths that learn automatically from the semantic paths of various relations.
Based on Section 3.3, we extend the multipath relation aggregation process to construct composite metapaths with two layers, as depicted in Figure 3. This method accounts for all possible interactions to construct the sub-metapaths and adjusts its weights according to the relative importance of the relations. Thus, when aggregating multiple neighboring nodes, this approach integrates the relations of nodes by considering the specific task scenarios through the learned weights.
More generally, the topological information of the l-layer metapath can be formulated as a combination of different unipath relation adjacency matrices, and the semantic information of the l t h -hop neighbor can be expressed as:
A ( l ) = A 1 A 2 A l
To flexibly and harmoniously adjust the relative importance of various relations in composite metapaths, it is necessary to calculate the adjacency matrices of the unipath networks representing these relations and then aggregate them based on their respective weights:
A ( 1 ) = i = 1 | R | w m i A i A ( 2 ) = i = 1 | R | j = 1 | R | w m i , m j A i A j A ( l ) = i = 1 | R | j = 1 | R | l w m i , , m j l A i A j l
where A ( l ) is the adjacency matrix of the l-layer composite metapaths. w m i , , m j represents the learned weights of different relations in the composite metapaths, and a total of i = 1 l | R | i learning parameters can be used to aggregate the diverse relations. The Conv operation fuses heterogeneous relations using trainable weights w m i , which are autonomously optimized through backpropagation to maximize task performance.
We assert that integrating the CMAE is essential, as it offers the following benefits to ANHNE:
  • A distinctive aspect of our approach is that relation weights do not require manual specification; rather, they are autonomously learned while considering the specific task scenarios. This allows for dynamic adjustment of weights based on the actual effects of embedding. For instance, in the academic network shown in Figure 1b, if the collaboration between authors holds more significance, the relations in the ‘APA’ path will carry higher weights. Conversely, if differences in research areas among authors are more important, the relations in the ‘APCPA’ path should have higher weights. This enables us to more precisely capture complex semantic information.
  • Our method enables automatic semantic mining. The composite metapaths learnd by CMAE are essentially soft combinations of various relations and can consider numerous sub-metapaths with specific semantics. Consequently, by multiplying and superimposing unipath relation adjacency matrices with varying weights, CMAE can autonomously extract l-layer metapaths, thereby overcoming the limitations of existing methods that require predefined metapaths. Unlike GTN [23], which generates metapaths via multiplication of adjacency matrices, our composite metapaths are weighted combinations of multipath relations.
  • We gather multi-layer relation topology information by multiplying matrices with weights, and then proceed to feature aggregation via GCN instead of stacking multi-layer GNNs. This not only reduces computational demands, but also avoids performance degradation due to over-smoothing.

3.5. Hierarchical Semantic Attention Aggregation Model (HSAA)

In many heterogeneous network representation methods, concentrating solely on the start and endpoint information of a metapath leads to incomplete features extraction. While merely linking the data from all nodes along the path offers a potential solution, it fails to effectively differentiate the varying roles of neighboring nodes at different hops in practical task scenarios, often resulting in redundancy.
To address this issue, this paper proposes a strategy for multi-layer neighbor information aggregation. First, based on the composite metapaths of different layers obtained in Section 3.4, the neighbor information of a specific layer is aggregated through one layer of GCN, as shown in Figure 4a. Then, the heterogeneous information of neighboring nodes of different layers is selectively aggregated according to the importance, as shown in Figure 4b. The HSAA model enables correct identification of neighbor semantics and accurate learning of network structure.
First, heterogeneous network data consists of nodes of various types with features residing in distinct semantic spaces. It is necessary to project attributes of nodes from different spaces to the same space. We define a projection matrix W T i , which maps X T i from the semantic space R d T i to the common semantic space R d C , and the projection formula is as follows:
H i = W T i · X T i
where X T i R V T i × d T i is the initial feature matrix of type T i , and W T i R d T i × d c is a type-specific projection matrix. For example, W P R d P × d c maps the feature matrix X P R V P × d P of the “Paper” type from the “Paper” space R d P to a new common space R d c .
Splice mapping feature matrices of diverse types to obtain the new feature mapping matrix of all nodes:
H = [ H 1 ; H 2 ; H T ]
To reduce computational complexity, we input the l-hop neighbor relation matrix A ( l ) , which contains multi-layer topology information, into a single-layer GCN to propagate node attributes:
Z ( l ) = σ ( D ( l ) 1 / 2 ( A ( l ) + I ) D ( l ) 1 / 2 H · W ( l ) )
where Z ( l ) R N × d is the aggregated information of the l-hop attributed neighbors propagating along the composite metapath, the i-th row Z i ( l ) of Z ( l ) denotes the semantic aggregated information of the node i, D ( l ) is the degree matrix of A ( l ) + I , W ( l ) R d c × d is the learnable matrix parameter, and σ is the activation function of ReLu, chosen for its sparsity induction to filter insignificant neighbor activations.
Neighbor nodes at different hops correspond to distinct semantics and properties. To effectively distinguish the importance of embedding expressions of these hop neighbors, we design an attention aggregation model. This model allows us to learn different hop-attention weights, evaluate the roles they actually play, and selectively aggregate the information of them. Specifically, the following equation is employed to determine the significance of various hops:
α l = σ δ ( l ) tanh W ( l ) Z ( l )
where α l represents the importance of l-hop neighbors’ information under the composite metapath, and δ ( l ) denotes the learnable matrix. Furthermore, the importance coefficients for the information at each hop are normalized using the softmax function:
μ l = exp α l j = 1 L exp α j
Finally, the aggregated information representations for various hops under the composite metapath are achieved:
Z = l = 1 L μ l Z ( l )
where Z is the synthesized information representation containing a combination of soft relations and can be used to perform various tasks.
In this paper, the cross-entropy function is employed as a mechanism to facilitate the effective feedback guidance role of labels on network embedding. The model’s loss can be computed as follows:
L r e ( Z , Y ) = i Ω j = 1 C Y i j log F Z i j
where C denotes the total count of node labels, Y represents the lable matrix containing 1 or 0 from the training set Ω , and F Z i j indicates the prediction result of the node i. Based on this loss function, the model is optimized using the gradient descent mechanism.
In conclusion, Equations (2)–(4) detail the construction of multi-length composite metapaths (RP1), upon which we efficiently aggregate the information across different hops using GCN, as demonstrated in Equations (7) and (10) (RP2).

3.6. Semantic Information Enhancement Model (SIE)

Although various combinations of explicit relations can be adaptively learned through l-layer composite metapaths, many nodes with similar homogeneous semantic information are still inevitably overlooked. These nodes are not connected through any kind of metapath, leading to a lack of implicit information. Consequently, in this section, to further explore implicit relations among nodes, we utilize the matured representation Z obtained in Section 3.5 to generate the potential network structure. This guides the acquisition of auxiliary node representation Z s as complementary information to learn a more robust embedded network and achieve node embeddings with superior performance (RP3).
First, to learn the potential graph topology, we measure the semantic relevance of two nodes using the pairwise cosine similarity of the two node representation vectors Z i and Z j , i.e., s i m ( Z i , Z j ) = < Z i , Z j > Z i Z j . The choice of K is motivated by a trade-off between preserving high-confidence semantic edges and suppressing noise: larger K increases computational cost and risks including irrelevant edges, while smaller K may overlook critical latent relations. This aligns with graph sparsification principles in prior work [48]. We choose the top-K semantically similar neighbors for each node to form the implicit adjacency matrix, as described below:
s i j K = 1 , s i j Top-K s i 0 , s i j Top-K s i
where s i j represents the i t h row and j t h column of the implicit neighbor information matrix S R N × N .
Subsequently, the features of its potential neighbors are captured using the implicit adjacency matrix:
Z s = σ ( D s 1 / 2 S K D s 1 / 2 H W s )
where D s is the degree matrix of the potential neighbor matrix S K and W s is the learnable matrix parameter.
Finally, the final node representation is generated by aggregation of the information representation Z and the auxiliary information representation Z s :
Z f = Z + l f Z s
where l f is a learning parameter that modulates the contribution of auxiliary information Z s to the final node representation.
Similarly, the loss in this module is calculated as:
L s e ( Z f , Y ) = i Ω j = 1 C Y i j log F Z f , i j
where Z f R N × d represents the final low-dimensional embedded feature representation.
The overall procedure is shown in Algorithm 1.
Electronics 14 02911 i001

3.7. Model Efficiency Analysis

ANHNE mainly consists of three modules: (1) composite metapath autonomous extraction (CMAE) involves matrix operations to construct multi-length metapaths. For a maximum path length L and relation types R , the complexity is O ( L · R L · N 2 ) due to weighted matrix multiplications; (2) hierarchical semantic attention aggregation (HSAA) projects heterogeneous features ( O ( N · d T i · d c ) ) and applies single-layer GCNs for each composite metapath ( O ( ε · d + N · d 2 ) ), with attention-based fusion across L hierarchical levels ( O ( L · N · d ) ); (3) semantic information enhancement (SIE) computes pairwise node similarities for implicit relations ( O ( N 2 · d ) but reduces this via top-K sparsification to O ( N log N + K · N · d ) , followed by a sparsity-aware GCN.
Overall, ANHNE avoids the high complexity of stacking multi-layer GNNs by leveraging CMAE for multi-hop relation capture and SIE’s sparsification. This design achieves efficiency comparable with state-of-the-art methods (e.g., GTN [23]) with only 1/10–1/100 computational overhead.

4. Experiments and Results Analysis

In this section, to demonstrate the superiority of the model presented in this paper, we undertake a series of experiments using publicly available datasets. Initially, we compare ANHNE against a range of state-of-the-art baseline methods in node classification and clustering tasks. Subsequently, we examine the practical utility of each component in enhancing network structure learning. Furthermore, we perform visualization experiments to assess the spatial distribution quality of the final node representations. Lastly, we conduct a hyperparameter study to investigate the influence of network parameter configurations on model performance.

4.1. Dataset

In our experiments, we utilize three widely recognized open-source heterogeneous network datasets commonly employed in the domain of network representation learning. Each dataset contains both attributes and relation information. The datasets are detailed as follows:
  • ACM (http://dl.acm.org (accessed on 3 July 2025)): This widely recognized open-source citation network dataset comprises nodes that represent papers, authors, and topics. The initial features of the papers are obtained through a bag-of-words model applied to the paper keywords, where each node is characterized by 1870 attributes.
  • DBLP (https://dblp.uni-trier.de (accessed on 3 July 2025)): This dataset features a citation network with nodes categorized into four types: papers, authors, conferences, and topics. For our experiments, we extracted a subset of the DBLP dataset, selecting 4057 nodes for analysis, which includes three types of paper nodes: information extraction, data mining, and artificial intelligence.
  • IMDB (https://www.imdb.com (accessed on 3 July 2025)): This dataset includes a citation network comprising nodes classified into four categories: papers, authors, conferences, and topics. For our experimental analysis, we selected a subset of the DBLP dataset, comprising 4057 nodes, which encompasses three types of paper nodes: information extraction, data mining, and artificial intelligence.
Further details are provided in Table 2.

4.2. Baselines

In these experiments, the model introduced in this paper was evaluated against a range of cutting-edge methods in the domain of heterogeneous network embedding. Table 3 offers a summary of these models.
  • GCN [10]: Extends traditional CNNs to non-Euclidean network structures, synthesizing neighbor node information at each layer through specific aggregation strategies (e.g., averaging or summing), and enabling in-depth exploration of the graph’s intrinsic properties and structural relations via feature analysis.
  • GAT [15]: Aggregates information from single-layer neighbors through an attention mechanism, allowing each node to assess the importance of its one-hop neighbors based on their features. However it can only aggregate information within one-hop neighbors.
  • HAN [16]: Learns the importance of nodes and their metapath-based neighbors, introducing semantic-level attention to fuse representations under different metapaths based on GAT, thus facilitating hierarchical feature aggregation through metapath neighbors.
  • MAGNN [49]: It projects attributes of heterogeneous nodes into a shared semantic space, then performs intra-metapath and inter-metapath information aggregation to capture the structural and semantic nuances of the network.
  • GTN [23]: Learns to softly select combined relations to generate useful metapaths and control lengths automatically. Its metapath generation module shares functionalities with the CMAE model in this study.
  • MHGCN [24]: Utilizes an attributed multi-order graph convolutional network to capture multi-relational structural information, explores different metapaths automatically, and employs both unsupervised and semi-supervised learning techniques to learn and derive the final node embeddings.
  • AMOGCN [45]: Constructs different order adjacency matrices containing various metapath relations, selectively fuses information from these matrices via SGC, and extracts final node embeddings through supervised learning that incorporates node semantic and labeling information.
  • ANHNE. Our proposed method.

4.3. Implementation Details

In the experiments, the ANHNE model is optimized using the Adam optimizer with a learning rate of 0.01, dropout rate of 0.5, and weight decay of 0.001. Other parameters are initialized using the Xavier distribution. The dimensionality of the common latent space after feature mapping is set to 128, while the final node embedding dimension is set to 64. The initial value of l f is set to 0.01. The length of composite metapaths L is searched within the range { 1 , 2 , 3 , 4 , 5 } , with the optimal value differing by dataset. Specifically, the optimal maximum path length L for the ACM dataset is chosen to be 2, whereas for the DBLP and IMDB datasets, it is selected as 3. An early stopping strategy is employed, with patience set to 15, to feed robust feature representations output from the HSAA module into the SIE module for implicit enhancement. We set K = 20 because at this point, the experimental performance of the three datasets has basically reached its optimum. For hyperparameter selection in all baseline methods, we adhere to the default settings from the original paper, which lead to optimal performance. For methods like GCN and HAN, we utilize a mean-value strategy to aggregate information from neighboring nodes of the target node. In methods such as MAGNN and HAN that require manually specified metapaths, we synthesize the settings from the original paper and use a uniform metapath. The node embedding dimension is consistently set to 64. To ensure result stability, all reported results are averaged over 10 experiments.
All experiments were conducted with the following hardware settings: NVIDIA GeForce RTX4060 GPU (Santa Clara, CA, USA), 16 GB × 2 RAM, Intel CPU i9-13900HX @2.20GHz 24-core, 2T KIOXIA SSD, running Win11 × 64. Our experiments were conducted using PyTorch 2.2.2, an open-source deep learning framework built on Python 3.11. The baseline models were implemented using the OpenHGNN library.

4.4. Performance on Embedding

Network training is initially guided by training set node labels to develop the final network embedding model. Subsequently, test set node feature vectors are fed into each model to perform downstream tasks, evaluating the overall performance of our proposed ANHNE and baseline models. Node classification, a supervised downstream task, and its results are presented in Table 4, while node clustering, an unsupervised task, is detailed in Table 5.
For node classification, we employ Micro-F1 and Macro-F1 scores to gauge the accuracy of each low-dimensional embedding results, offering a balanced and comprehensive assessment of node embedding quality. According to Table 4, ANHNE consistently outperforms in both metrics across various training set proportions. The lower performance of GCN and GAT, typically suited for homogeneous networks, is attributed to their inadequacy in discriminating among the heterogeneous dataset’s diverse node types and complex relations. GAT’s superior performance over GCN suggests that its attention mechanism effectively aids in learning better network embeddings. In contrast, other heterogeneous network methods surpass isomorphic ones, integrating a wider array of node features and network structures. Both MHGCN and AMOGCN lead the baseline models, indicating that multi-stage graph convolution effectively prevents overfitting and achieves high-performance network embeddings. Compared with the best results of the baseline model, ANHEN’s Macro-F1 improves by 1.28%, 0.92%, and 4.90% across three datasets, and Micro-F1 improves by 1.18%, 0.79%, and 0.66%. This demonstrates that ANHEN can achieve effective embedding for node classification tasks. Further assessing model performance, we conduct unsupervised downstream node clustering experiments using a k-means model, setting the number of clusters to match the node category count in each dataset. Employing NMI and ARI as clustering metrics, Table 5 reveals ANHNE’s superiority over all baseline models in node clustering tasks, with average improvements of 2.00% in NMI and 3.31% in ARI scores compared with the best baseline results. Similar to node classification outcomes, heterogeneous network models generally outperform homogeneous ones, and multi-order graph convolution models excel over other baselines.
In summary, the results affirm ANHNE’s effectiveness in node classification and clustering tasks, highlighting its superior node representation learning and general applicability across various downstream tasks.

4.5. Ablation Experiment

ANHNE integrates HIN embedding through the CMAE, HSAA, and SIE modules. To better assess the functionality of each component, we conduct ablation studies to evaluate the impact of various model variants on embedding performance, specifically:
  • Only manual metapath aggregation (OMMA), which replaces composite metapaths with manual and fixed ones. This variant aggregates information of multi-hop nodes across common manually defined metapaths instead of composite metapaths with diverse sub-metapaths.
  • No layer-level attention aggregation (NOAA), which omits the HSAA module. This variant does not differentiate the semantics of different neighbor layers but aggregates all layers of neighborhood information equally.
  • No semantic information enhancement (NOSIE), which overlooks potential inter-node information. This variant discards the consideration of node homogeneity and relies solely on the embedding learned through HSAA outlined in Section 3.5 as the final network representation.
We performed ablation studies using the DBLP dataset, maintaining the same structural framework as ANHNE across three model variants, except for the particular alterations specified. The embedding performance of these variants and ANHNE is depicted in Figure 5.
Analysis of the results shows that all three variants exhibit reduced performance compared with ANHNE. NOSIE demonstrates the most significant performance drop, which is attributable to its disregard for latent node information, focusing solely on explicit relations for feature aggregation. The significant performance gap between ANHNE and NOSIE arises because NOSIE discards latent node correlations derived from attribute homophily. Whereas explicit metapaths encode direct relations, implicit links (e.g., technique-sharing papers P1/P4 in Figure 1b) rely on attribute-based affinities. Ignoring such semantics weakens representation inclusivity, evidenced by NOSIE’s classification clustering degradation. OMMA underperforms compared with ANHNE, indicating that the metapath autonomous extraction module’s ability to derive useful sub-metapath combinations and effectively aggregate features is crucial. This module allows for adjusting the proportions of different sub-metapaths according to the task scenario, enhancing the interaction of effective features and achieving a high-performing network embedding representation. Among the variants, NOAA often shows the best performance, suggesting that the hierarchical attention mechanism proficiently discriminates between the attribute semantics and feature representations of different hierarchical neighbors, selectively aggregating this information to improve generalization capabilities and positively influence embedding outcomes. In summary, our ablation studies confirm the positive contributions of each component to node embedding.

4.6. Analysis Experiment

In this section, we further conduct several experiments necessary to seek a deeper understanding of our ANHNE approach. First, we conduct experiments to demonstrate the effectiveness of the automatic learning of combinatorial metapaths by identifying the learned metapaths. Second, it is verified that the ANHNE method has discriminative aggregation for different layers of neighborhood information. Finally, the nodes represent an intuitive visualization of the discrimination.
(1)
Do the composite metapaths learned by the CMAE module enable automatic semantic mining? We recorded the fusion weights of different relations in the CMAE module. We multiply the weights of different unipath relations with the hierarchical attention, set them to the weights of different sub-metapath relations, and select the paths with the top three weights. As can be seen from Table 6, our approach not only learns the commonly used metapaths in various domains, which are often used as a priori information for metapath-based models (e.g., HAN), but also further mines metapaths containing rich information to enhance the representation of nodes. For instance, the ‘APAPA’ metapath in DBLP, which represents two authors who have collaborated with the same author on different papers, can be mined for higher-order collaborations. Authors involved in such higher-order collaborations are more likely to be part of the same research field, significantly enhancing the performance of node classification and clustering.
Overall, the CMAE module is able to extract sub-metapaths containing different semantic information, and can autonomously regulate their weights to enhance semantic information aggregation. Meanwhile, the experimental results in Section 4.5 show that the automatic discovery of composite metapaths is more conducive to embedding performance than defining metapaths manually.
(2)
Does our proposed HSAA module effectively distinguish the semantic features of different hopping neighbors? To confirm the efficacy of integrating the information from nodes within the composite metapath, we conducted experiments on the relation between the attention weights of the neighbor information of different layers and the clustering performance of the corresponding layers. Figure 6 illustrates the attention weights for single-layer information alongside the clustering performance of the respective layer. Due to space limitations, we display results only for the ACM and IMDB datasets; however, the DBLP dataset exhibits an identical trend. There is a conspicuous positive correlation between the attention weights and clustering performance. This demonstrates that the model effectively prioritizes the neighborhood information of critical layers, substantially enhancing the overall performance and confirming the efficacy of the HSAA module introduced in Section 3.5. On the other hand, the intermediate layer node information also has higher attention and contribution to the model performance, proving the importance of the metapath internal node information. It is in line with the initial hypothesis that the metapath internal node information can help to accurately mine the complex relationship between the start and destination nodes and improve the model performance.
(3)
Does our proposed ANHNE model effectively enhance the distinction among network representations? We carried out visualization experiments to evaluate the quality of these representations, a standard method for assessing network representation tasks. Employing T-SNE, we compressed the node feature embeddings into a two-dimensional space to benchmark our model against leading techniques in the field. Figure 7 illustrates the distribution of nodes for various models on the DBLP dataset, where different colors denote different node types. In conjunction with the findings presented in Section 4.4, we contend that the low-dimensional feature space distributions generated by all approaches accurately demonstrate the relationships with node classes. Obviously, the homogeneous model GCN has the worst performance, where different types of nodes are mixed together and clusters of nodes do not form clear boundaries. HAN improves on it to a certain extent, but the improvement is not obvious; although the red clusters are clearly separated, the yellow clusters of nodes are collocated with the green and purple clusters of nodes and do not form clear boundaries. In comparison, the performance of AMOGCN is significantly improved, with the node clusters of all four colors forming their own compact regions; however, unfortunately, it has various types of nodes intertwined in the middle region. Compared with the baseline model, our model more effectively compacts the spatial distribution of nodes with identical labels and minimizes the cross-over between nodes of different categories, resulting in a more distinct boundary. This indicates that our model excels in differentiating between various node types, possesses enhanced generalization capabilities, and offers robust support for downstream tasks.

4.7. Hyperparameter Study

In our ANHNE model, we examine the sensitivity by adjusting hyperparameters and observing the resultant variations in node classification and clustering performance. Sensitivity tests conducted on the DBLP dataset demonstrate consistent trends across the ACM and IMDB datasets.
Sensitivity of K, the number of semantically similar nodes: Within our semantic information enhancement module, the parameter K dictates the depth of complementary hidden layer semantics. To assess the impact of various semantic depths on downstream task performance, we explore a range of K values: { 0 , 10 , 30 , 60 , 100 , 150 } . Figure 8 depicts the embedding performance trend as K increases. Initially, both classification and clustering performances improve with increasing depth of supplementary information, suggesting that a higher K value can unearth richer hidden information and consequently enhance the network structure. However, beyond a K value of 20, performance diminishes significantly, indicative of excessive hidden information leading to long path redundancies and noise introduction, which in turn degrades embedding performance.
Sensitivity of d, the embedding dimension: Different embedding dimensions capture varying degrees of feature and structural information. To investigate the effect of embedding dimensionality on model performance, we set dimensions ranging { 16 , 32 , 64 , 128 , 256 , 512 } . Figure 8 shows the classification and clustering performance trends, respectively. Initially, performance enhances as all four measures increase with dimensionality, implying that higher dimensional representations capture feature and semantic information more comprehensively. However, at a dimensionality of 512, both classification and clustering performances decline slightly, suggesting that excessively high-dimensional node embeddings introduce noise that can skew the network structure.

4.8. Convergence Analysis

To demonstrate the good convergence of ANHNE, this experiment presents the trends of evaluation metrics and loss values across three test datasets. As shown in Figure 9, accuracy rapidly increases with training iterations and exhibits slight fluctuations after reaching its optimal state, while loss values decrease sharply across all datasets and stably converge within a narrow range above and below a specific value after sufficient iterations. These observations confirm the strong convergence capability of ANHNE.

5. Conclusions

This paper introduces a network embedding model, ANHNE, designed for heterogeneous networks, with three main contributions: First, we expand the multipath relation aggregation approach and design composite metapaths that can autonomously extract diverse lengths containing complex relations. This allows for the automatic mining of deep semantics according to specific task scenarios, thereby addressing the limitations of existing methods that require predefined metapaths. Second, we map the features of different node types to a public semantic space and, using attention coefficients, quantify the importance of various levels of semantic information on composite metapaths to fully leverage neighborhood information. Third, we introduce a semantic information enhancement module to explore implicit node relations, compensating for semantic deficits caused by overlooking node homogeneity and aggregating more implicit features.
Experimental results affirm that ANHNE achieves top-tier or competitive node embedding performance and can autonomously learn valuable composite metapaths based on task scenarios. Currently implemented on a static network, ANHNE’s adaptive metapath learning and multi-hop fusion architecture offer a potential pathway for handling dynamic heterogeneous networks. The semantic attention mechanism, which selectively aggregates neighborhood hierarchies, could be extended to model-evolving node–edge interactions. Future work will deploy ANHNE on large-scale dynamic graphs to refine its scalability, computational efficiency, and real-time adaptability.

Author Contributions

H.X.: Conceptualization, data curation, formal analysis, funding acquisition, investigation, methodology, project administration, software, validation, visualization, writing—original draft, writing—review and editing; H.S.: Formal analysis, investigation, methodology, validation, writing—original draft, writing—review and editing; L.W.: Data curation, supervision, formal analysis, resources, validation, writing—original draft, writing—review and editing; C.S.: Formal analysis, funding acquisition, investigation, methodology, resources, supervision, validation, writing—original draft, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

The authors express their gratitude for the financial support provided by the National Science Foundation of China (62301581) and partially provided by the Science Foundation of Anhui Province (2008085QF326).

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, D.; Yin, J.; Zhu, X.; Zhang, C. Network representation learning: A survey. IEEE Trans. Big Data 2018, 6, 3–28. [Google Scholar] [CrossRef]
  2. Taskar, B.; Wong, M.F.; Abbeel, P.; Koller, D. Link prediction in relational data. Adv. Neural Inf. Process. Syst. 2003, 16, 1–8. [Google Scholar]
  3. Bhagat, S.; Cormode, G.; Muthukrishnan, S. Node Classification in Social Networks; Springer: Berlin/Heidelberg, Germany, 2011; pp. 115–148. [Google Scholar]
  4. Fan, W.; Ma, Y.; Li, Q.; He, Y.; Zhao, E.; Tang, J.; Yin, D. Graph neural networks for social recommendation. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 417–426. [Google Scholar]
  5. Qiu, J.; Tang, J.; Ma, H.; Dong, Y.; Wang, K.; Tang, J. Deepinf: Social influence prediction with deep learning. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 2110–2119. [Google Scholar]
  6. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  7. Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 1067–1077. [Google Scholar]
  8. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery And Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
  9. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  10. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  11. Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional neural networks on graphs with fast localized spectral filtering. arXiv 2016, arXiv:1606.09375. [Google Scholar]
  12. Atwood, J.; Towsley, D. Diffusion-convolutional neural networks. arXiv 2016, arXiv:1511.02136. [Google Scholar]
  13. Chen, Y.; Chen, F.; Wu, Z.; Chen, Z.; Cai, Z.; Tan, Y.; Wang, S. Heterogeneous Graph Embedding with Dual Edge Differentiation. Neural Netw. 2025, 183, 106965. [Google Scholar] [CrossRef]
  14. Chen, Y.; Song, A.; Yin, H.; Zhong, S.; Chen, F.; Xu, Q.; Wang, S.; Xu, M. Multi-view incremental learning with structured hebbian plasticity for enhanced fusion efficiency. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; Volume 39, pp. 1265–1273. [Google Scholar]
  15. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. In Proceedings of the ICLR 2018 Conference Track 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  16. Wang, X.; Ji, H.; Shi, C.; Wang, B.; Ye, Y.; Cui, P.; Yu, P.S. Heterogeneous graph attention network. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2022–2032. [Google Scholar]
  17. Park, C.; Han, J.; Yu, H. Deep multiplex graph infomax: Attentive multiplex network embedding using global information. Knowl.-Based Syst. 2020, 197, 105861. [Google Scholar] [CrossRef]
  18. Jing, B.; Park, C.; Tong, H. Hdmi: High-order deep multiplex infomax. In Proceedings of the Web Conference 2021, Ljubljana, Slovenia, 19–23 April 2021; pp. 2414–2424. [Google Scholar]
  19. Zhao, M.; Yu, J.; Zhang, S.; Jia, A.L. Relation-aware multiplex heterogeneous graph neural network. Knowl.-Based Syst. 2025, 309, 112806. [Google Scholar] [CrossRef]
  20. Su, H.; Li, Q.; Gong, Y.; Liu, Y.; Jiang, X. Attribute Disturbance for Attributed Multiplex Heterogeneous Network Embedding. In Proceedings of the 2024 5th International Conference on Big Data & Artificial Intelligence & Software Engineering (ICBASE), Wenzhou, China, 20–22 September 2024; pp. 455–462. [Google Scholar]
  21. Ding, L.; Li, M.; Wang, Y.; Shi, P.; Zhang, F. AMIRLe: Attribute-enhanced multi-interaction representation learning fore-commerce heterogeneous information networks. Int. J. Mach. Learn. Cybern. 2024, preprint. [Google Scholar] [CrossRef]
  22. Cen, Y.; Zou, X.; Zhang, J.; Yang, H.; Zhou, J.; Tang, J. Representation learning for attributed multiplex heterogeneous network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 1358–1368. [Google Scholar]
  23. Yun, S.; Jeong, M.; Kim, R.; Kang, J.; Kim, H.J. Graph transformer networks. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
  24. Yu, P.; Fu, C.; Yu, Y.; Huang, C.; Zhao, Z.; Dong, J. Multiplex heterogeneous graph convolutional network. In Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 14–18 August 2022; pp. 2377–2387. [Google Scholar]
  25. Dong, Y.; Chawla, N.V.; Swami, A. metapath2vec: Scalable representation learning for heterogeneous networks. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; pp. 135–144. [Google Scholar]
  26. Hu, Z.; Dong, Y.; Wang, K.; Sun, Y. Heterogeneous graph transformer. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 2704–2710. [Google Scholar]
  27. Jin, B.; Gao, C.; He, X.; Jin, D.; Li, Y. Multi-behavior recommendation with graph convolutional networks. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 25–30 July 2020; pp. 659–668. [Google Scholar]
  28. Qin, X.; Sheikh, N.; Reinwald, B.; Wu, L. Relation-aware graph attention model with adaptive self-adversarial training. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 9368–9376. [Google Scholar]
  29. Etta, G.; Cinelli, M.; Galeazzi, A.; Valensise, C.M.; Quattrociocchi, W.; Conti, M. Comparing the impact of social media regulations on news consumption. IEEE Trans. Comput. Soc. Syst. 2022, 10, 1252–1262. [Google Scholar] [CrossRef]
  30. Ren, Y.; Liu, B.; Huang, C.; Dai, P.; Bo, L.; Zhang, J. Heterogeneous deep graph infomax. arXiv 2019, arXiv:1911.08538. [Google Scholar]
  31. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef]
  32. Manchanda, S.; Zheng, D.; Karypis, G. Schema-aware deep graph convolutional networks for heterogeneous graphs. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data), Orlando, FL, USA, 15–18 December 2021; pp. 480–489. [Google Scholar]
  33. Wang, X.; Zhu, M.; Bo, D.; Cui, P.; Shi, C.; Pei, J. Am-gcn: Adaptive multi-channel graph convolutional networks. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual, 6–10 July 2020; pp. 1243–1253. [Google Scholar]
  34. Min, Y.; Wenkel, F.; Wolf, G. Scattering gcn: Overcoming oversmoothness in graph convolutional networks. Adv. Neural. Inf. Process. Syst. 2020, 33, 14498–14508. [Google Scholar] [PubMed]
  35. Huang, W.; Hao, F.; Shang, J.; Yu, W.; Zeng, S.; Bisogni, C.; Loia, V. Dual-LightGCN: Dual light graph convolutional network for discriminative recommendation. Comput. Commun. 2023, 204, 89–100. [Google Scholar] [CrossRef]
  36. Feng, F.; He, X.; Zhang, H.; Chua, T.S. Cross-GCN: Enhancing Graph Convolutional Network with k-Order Feature Interactions. IEEE Trans. Knowl. Data Eng. 2021, 35, 225–236. [Google Scholar] [CrossRef]
  37. Wu, F.; Souza, A.; Zhang, T.; Fifty, C.; Yu, T.; Weinberger, K. Simplifying graph convolutional networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6861–6871. [Google Scholar]
  38. Zhang, X.; Liu, H.; Li, Q.; Wu, X.M. Attributed graph clustering via adaptive graph convolution. arXiv 2019, arXiv:1906.01210. [Google Scholar]
  39. Wang, X.; Liu, N.; Han, H.; Shi, C. Self-supervised Heterogeneous Graph Neural Network with Co-contrastive Learning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 14–18 August 2021; pp. 1726–1736. [Google Scholar] [CrossRef]
  40. Park, C.; Kim, D.; Han, J.; Yu, H. Unsupervised attributed multiplex network embedding. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 5371–5378. [Google Scholar]
  41. Fang, Y.; Zhao, X.; Chen, Y.; Xiao, W.; de Rijke, M. PF-HIN: Pre-Training for Heterogeneous Information Networks. IEEE Trans. Knowl. Data Eng. 2022, 35, 8372–8385. [Google Scholar]
  42. Yuan, R.; Wu, Y.; Tang, Y.; Wang, J.; Zhang, W. Meta-path infomax joint structure enhancement for multiplex network representation learning. Knowl.-Based Syst. 2023, 275, 110701. [Google Scholar] [CrossRef]
  43. Shi, C.; Hu, B.; Zhao, W.X.; Philip, S.Y. Heterogeneous information network embedding for recommendation. IEEE Trans. Knowl. Data Eng. 2018, 31, 357–370. [Google Scholar] [CrossRef]
  44. Mei, G.; Pan, L.; Liu, S. Heterogeneous graph embedding by aggregating meta-path and meta-structure through attention mechanism. Neurocomputing 2022, 468, 276–285. [Google Scholar] [CrossRef]
  45. Chen, Z.; Wu, Z.; Zhong, L.; Plant, C.; Wang, S.; Guo, W. Attributed Multi-order Graph Convolutional Network for Heterogeneous Graphs. Neural Netw. 2024, 174, 106225. [Google Scholar] [CrossRef]
  46. Xue, H.; Sun, X.K.; Sun, W.X. Multi-hop hierarchical graph neural networks. In Proceedings of the 2020 IEEE International Conference on Big Data and Smart Computing (BigComp), Busan, Republic of Korea, 19–22 February 2020; pp. 82–89. [Google Scholar]
  47. Sun, Y.; Zhu, D.; Du, H.; Tian, Z. MHNF: Multi-hop heterogeneous neighborhood information fusion graph representation learning. IEEE Trans. Knowl. Data Eng. 2022, 35, 7192–7205. [Google Scholar] [CrossRef]
  48. Zhu, Y.; Xu, W.; Zhang, J.; Du, Y.; Zhang, J.; Liu, Q.; Yang, C.; Wu, S. A survey on graph structure learning: Progress and opportunities. arXiv 2021, arXiv:2103.03036. [Google Scholar]
  49. Fu, X.; Zhang, J.; Meng, Z.; King, I. Magnn: Metapath aggregated graph neural network for heterogeneous graph embedding. In Proceedings of the Web Conference 2020, Taipei, Taiwan, 20–24 April 2020; pp. 2331–2341. [Google Scholar]
Figure 1. A partial heterogeneous network DBLP for an illustration of key challenges. (a) Three node types in the DBLP. (b) Its attributed multiplex structure with four edge types (e.g., author–paper and paper–conference). (c) Metapaths APA (author–paper–author) and PCP (paper–conference–paper) reflecting collaboration and venue-based semantics, and (d) a unipath relation network.
Figure 1. A partial heterogeneous network DBLP for an illustration of key challenges. (a) Three node types in the DBLP. (b) Its attributed multiplex structure with four edge types (e.g., author–paper and paper–conference). (c) Metapaths APA (author–paper–author) and PCP (paper–conference–paper) reflecting collaboration and venue-based semantics, and (d) a unipath relation network.
Electronics 14 02911 g001
Figure 3. Illustration of metapaths with importance for a toy example.
Figure 3. Illustration of metapaths with importance for a toy example.
Electronics 14 02911 g003
Figure 4. (a) The information aggregation process of one layer in a composite metapath. (b) The information aggregation process across layers in a composite metapath with importance weights.
Figure 4. (a) The information aggregation process of one layer in a composite metapath. (b) The information aggregation process across layers in a composite metapath with importance weights.
Electronics 14 02911 g004
Figure 5. Embedding performance of model variants. Performance trends align across metrics as ablation of key components homogeneously degrades classification (Macro/Micro-F1) and clustering (NMI/ARI) quality.
Figure 5. Embedding performance of model variants. Performance trends align across metrics as ablation of key components homogeneously degrades classification (Macro/Micro-F1) and clustering (NMI/ARI) quality.
Electronics 14 02911 g005
Figure 6. Attention weights of different layers and clustering performance of corresponding layers in the ACM and IMDB datasets.
Figure 6. Attention weights of different layers and clustering performance of corresponding layers in the ACM and IMDB datasets.
Electronics 14 02911 g006
Figure 7. T-SNE space distribution diagram for node embeddings of compared models. Each colored node represents the category label of that particular color.
Figure 7. T-SNE space distribution diagram for node embeddings of compared models. Each colored node represents the category label of that particular color.
Electronics 14 02911 g007
Figure 8. Parameter sensitivity of hyperparameters K and d in ANHNE.
Figure 8. Parameter sensitivity of hyperparameters K and d in ANHNE.
Electronics 14 02911 g008
Figure 9. Convergence curves of loss values (purple), training accuracy (black), and validation accuracy (red) of ANHNE on the tested datasets.
Figure 9. Convergence curves of loss values (purple), training accuracy (black), and validation accuracy (red) of ANHNE on the tested datasets.
Electronics 14 02911 g009
Table 1. A summary of primary notations in this paper.
Table 1. A summary of primary notations in this paper.
NotationsExplanations
I R N × N The identity matrix.
σ The activation function.
T i The type of node i.
R r The r-th node relation.
X T i The initial feature matrix of the node of type T i .
H i The projected node feature of type T i .
A r The adjacency matrix under the relation R r .
A ( l ) The adjacency matrix of the l-layer composite metapath.
W T i The feature mapping matrix for the type T i .
W ( l ) The trainable weight matrix of l-layer GCN.
β i The weight of relation i in the composite metapath.
α l The composite metapath-based hierarchical attention weight for l-layer neighbors.
l f The learning parameter to modulate the contribution of auxiliary information.
S K The latent graph structure adjacency matrix.
Z ( l ) The ndoe embedding of composite metapath based l-layer neighbors.
Z The node embedding of multi-layer neighbors.
Z s The latent graph structure node embedding.
Z f The final node embedding.
Y Ground truth.
Table 2. Datasets of heterogeneous networks.
Table 2. Datasets of heterogeneous networks.
DatasetsNum.NodesAttributesNode TypesClassesTrain/Val/Test
ACMPaper(3025)1870Paper(P)/Author(A)/Subject(S)3600/300/2125
DBLPPaper(4057)334Paper(P)/Author(A)/Conference(C)/Term(T)4800/400/2857
IMDBMovie(3550)1232Movie(M)/Actor(A)/Director(D)/Year(Y)3600/300/2660
Table 3. Characteristics of different models (Attri.: attributed; Hetero.: for heterogeneous networks; Learnable.: learnable metapaths; Multi-order.: multi-order metapaths).
Table 3. Characteristics of different models (Attri.: attributed; Hetero.: for heterogeneous networks; Learnable.: learnable metapaths; Multi-order.: multi-order metapaths).
MethodsAttri.Hetero.Learnable.Multi-Order.
GCN×××
GAT×××
HAN××
MAGNN××
GTN××
MHGCN
AMOGCN
ANHNE
Table 4. Quantitative results on the node classification task (%). Bold indicates “the best”, and underline indicates “the second best”.
Table 4. Quantitative results on the node classification task (%). Bold indicates “the best”, and underline indicates “the second best”.
DatasetsMetricsTrainingGCNGATHANMAGNNGTNMHGCNAMOGCNANHNE
ACMMacro-F120%86.9287.8787.9490.7992.5792.6492.1893.78
40%87.7388.4488.6890.9692.5392.5692.3794.01
60%87.8689.2589.1591.0392.8493.2892.2894.11
80%88.0889.5589.8091.0892.6592.6992.3094.02
Micro-F120%86.7287.6987.6790.6392.4492.5692.0793.71
40%87.5588.2888.4290.7892.5592.6192.4094.04
60%87.6989.1288.9290.8692.5693.3892.3894.05
80%87.9489.4389.6090.8992.9192.7292.3894.04
DBLPMacro-F120%90.9690.0591.6692.0290.4691.9992.3193.33
40%91.3791.2091.8892.1790.6992.2392.9193.80
60%91.6191.3592.0992.2090.9992.6592.9093.86
80%91.8691.4492.1092.1790.9393.1193.4293.97
Micro-F120%91.8491.7092.6192.9291.0892.4893.1093.60
40%92.1892.1592.7893.0691.2892.7393.3494.09
60%92.4392.3193.0093.1091.5693.1093.1194.33
80%92.6692.3793.0793.0691.4993.3393.5894.07
IMDBMacro-F120%45.7349.4450.0051.9851.1352.1751.0753.43
40%48.0150.6452.7152.5552.0753.6452.6557.67
60%49.1551.9054.2454.1154.2954.8453.6257.32
80%51.8152.9954.3854.5954.6853.8652.9957.49
Micro-F120%49.7855.2859.1660.7760.1062.8163.0763.18
40%51.7155.9160.8361.3760.3265.0664.6465.48
60%52.2956.4462.3561.7760.3366.1266.6566.74
80%54.1656.9763.4462.7660.3366.9566.1167.36
Table 5. Quantitative results on the node clustering task (%). Bold indicates “the best”, and underline indicates “the second best”.
Table 5. Quantitative results on the node clustering task (%). Bold indicates “the best”, and underline indicates “the second best”.
DatasetsMetricsGCNGATHANMAGNNGTNMHGCNAMOGCNANHNE
ACMNMI58.7863.1966.4972.0374.9275.8372.5478.09
ARI62.6567.7570.5676.5679.8081.5978.8883.05
DBLPNMI71.5574.2275.4977.0177.2779.3079.8881.25
ARI76.3179.4381.3281.3982.1083.4283.9185.01
IMDBNMI9.5910.0213.0815.5917.6916.8317.0917.92
ARI6.598.6910.9413.3618.6823.6722.6925.29
Table 6. The common metapath and the top three metapaths identified across various datasets.
Table 6. The common metapath and the top three metapaths identified across various datasets.
DatasetsCommon MetapathLearned Metapath
ACMPSP, PAPPAP, APA, PSP
DBLPAPA, APAPAAPA, APAPA, APCPA
IMDBMDM, MAMMAM, AMA, MDM
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, H.; Shao, H.; Wang, L.; Song, C. ANHNE: Adaptive Multi-Hop Neighborhood Information Fusion for Heterogeneous Network Embedding. Electronics 2025, 14, 2911. https://doi.org/10.3390/electronics14142911

AMA Style

Xie H, Shao H, Wang L, Song C. ANHNE: Adaptive Multi-Hop Neighborhood Information Fusion for Heterogeneous Network Embedding. Electronics. 2025; 14(14):2911. https://doi.org/10.3390/electronics14142911

Chicago/Turabian Style

Xie, Hanyu, Hao Shao, Lunwen Wang, and Changjian Song. 2025. "ANHNE: Adaptive Multi-Hop Neighborhood Information Fusion for Heterogeneous Network Embedding" Electronics 14, no. 14: 2911. https://doi.org/10.3390/electronics14142911

APA Style

Xie, H., Shao, H., Wang, L., & Song, C. (2025). ANHNE: Adaptive Multi-Hop Neighborhood Information Fusion for Heterogeneous Network Embedding. Electronics, 14(14), 2911. https://doi.org/10.3390/electronics14142911

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop