Next Article in Journal
Goalie: Defending Against Correlated Value and Sign Encoding Attacks
Previous Article in Journal
Weibull-Type Incubation Period and Time of Exposure Using γ-Divergence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DeeWaNA: An Unsupervised Network Representation Learning Framework Integrating Deepwalk and Neighborhood Aggregation for Node Classification

1
School of Media Science, Northeast Normal University, Jingye Street 2555, Changchun 130117, China
2
School of Journalism, Northeast Normal University, Jingye Street 2555, Changchun 130117, China
3
School of Information Science and Technology, Northeast Normal University, Jingye Street 2555, Changchun 130117, China
4
School of Physics, Northeast Normal University, Renmin Street 5268, Changchun 130024, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(3), 322; https://doi.org/10.3390/e27030322
Submission received: 6 February 2025 / Revised: 3 March 2025 / Accepted: 11 March 2025 / Published: 20 March 2025

Abstract

:
This paper introduces DeeWaNA, an unsupervised network representation learning framework that unifies random walk strategies and neighborhood aggregation mechanisms to improve node classification performance. Unlike existing methods that treat these two paradigms separately, our approach integrates them into a cohesive model, addressing limitations in structural feature extraction and neighborhood relationship modeling. DeeWaNA first leverages DeepWalk to capture global structural information and then employs an attention-based weighting mechanism to refine neighborhood relationships through a novel distance metric. Finally, a weighted aggregation operator fuses these representations into a unified low-dimensional space. By bridging the gap between random-walk-based and neural-network-based techniques, our framework enhances representation quality and improves classification accuracy. Extensive evaluations on real-world networks demonstrate that DeeWaNA outperforms four widely used unsupervised network representation learning methods, underscoring its effectiveness and broader applicability.

1. Introduction

Network Representation Learning (NRL) aims to learn latent, low-dimensional network representational vectors [1]. With information networks becoming ubiquitous in describing the complex relationships in large real-world applications in the forms of social networks, biological networks, and citation networks, NRL has played a crucial role in many network analysis applications, including node classification, link prediction, community detection, and subnetwork similarity [1,2,3,4]. Because of the advantages of adaption for the absence of expensive manual annotation data, many Unsupervised Network Representation Learning (UNRL) methods [5,6,7,8,9,10] have been proposed and have shown remarkable achievements.
DeepWalk is one of the widely used random-walk-based UNRL algorithms, due to its few-shot friendliness and simple distributed computation. It is based on context sampling of node sequences on networks to calculate representation vectors, which results in a reduced capacity for capturing neighborhood messages. Many researches have demonstrated the importance of catching neighborhood messages in network representation learning, and various neighborhood aggregation methods have been proposed [11,12,13,14]. Integrating DeepWalk with neighborhood aggregation techniques appears to be a promising strategy that provides representational vectors with improved performance for machine learning models in downstream network analysis tasks.
Although the existing aggregation methods differ from model to model [15], the distances between nodes and their neighbors in embedding space are not taken into account by the approaches. The updated representational vectors may cause tasks to become more challenging when neighborhood nodes that are located at varying distances from a central node are considered equally throughout neighborhood aggregation operators. We illustrate an example of a node classification task, as shown in Figure 1. Given a network G, as shown in Figure 1a, neighborhood aggregation techniques can reduce the difficulty of the classification task from the perspective of position in the embedding space, as shown in Figure 1b. In certain circumstances, poor aggregation techniques will arrange neighbor nodes closely together in the embedding space, as shown in Figure 1c. In some awful circumstances, catastrophic aggregation techniques might make tasks more difficult, as shown in Figure 1d. As a result, the neighbors should be distinguished by distinct weights throughout the aggregation process. Many works have used the attention mechanism, a weighting technique, to discriminate between neighbors [10,16].
In this paper, we presented the DeeWaNA framework for unsupervised network representation learning, which combines DeepWalk with neighborhood aggregation. Specifically, DeeWaNA initially makes use of DeepWalk to obtain the nodes’ self-representational vectors. Then, the framework employs a weighted neighborhood aggregation operation to produce local representational vectors of nodes based on their self-representational vectors. To distinguish between the various effects of neighbor nodes on their center node, DeeWaNA uses the self-attention between nodes and their one-hop neighbor nodes as the weights in the weighted neighborhood aggregation process. For the purpose of obtaining self-attention, we suggest a Euclidean distance-based learning task on representational vectors in embedding space. DeeWaNA fuses the sequential connecting self-representational vectors and local representational vectors into a single unified embedding space using the Principal Component Analysis (PCA) method [17].
We conducted thorough tests to evaluate DeWaNA against four widely used unsupervised network representation learning algorithms on a range of four real-world datasets for the single-label/multi-label node classification tasks. Our experiments demonstrate that DeeWaNA can provide the network representational with enhanced performance to machine learning models on single-label/multi-label node classification tasks.

2. Related Work

It is common practice to employ random-walk-based approaches for unsupervised network representation learning [18,19]. Deepwalk is based on a strategy of choosing neighbors with equal probability for sampling node sequences that are fed into the skip-gram model for network embedding [6]. The node2vec algorithm is an improved version of DeepWalk, which utilizes two parameters for adjusting the usage of a deep-first search and a breadth-first search in sampling node sequences [9]. The off-the-shelf random-walk-based approach, however, is unable to fully utilize the neighbors’ message that is frequently employed in deep-structure-based network representation learning models.
Many deep-structure-based models have been proposed for network representation learning [5,11,12]. Neighborhood aggregation plays an important role in neural-network-based models for learning network representational vectors with high performance [5,13,15,20,21,22]. These studies do not take into account the various effects of neighbors on aggregation operators. By utilizing masked self-attentional stacking layers, which may implicitly assign various weights to different nodes in a neighborhood without expensive matrix operation or knowing graph topology beforehand, Graph Attention Networks (GATs) [10] are utilized to attend to the impacts of neighbors. Although GATs can discriminate between various neighbors, they need labels to acquire self-attention meaning; hence, designing learning tasks is necessary when employing GATs in unsupervised network representation learning.

3. Methodology

In this section, we offer a framework called DeeWaNA that combines DeepWalk with weighted neighborhood aggregation for unsupervised network representation learning. DeeWaNA incorporates weighted neighborhood aggregation into DeepWalk to improve previous random-walk-based unsupervised network representation learning techniques that are unable to collect neighbor information. Consequently, by utilizing DeepWalk and the weighted neighborhood aggregation technique to gather neighborhood information, DeeWaNA can produce high-performance network representation.

3.1. The General Framework of DeeWaNA

We first present the general framework of DeeWaNA, which incorporates DeepWalk and weighted neighborhood aggregation techniques and is illustrated in Figure 2.
First, given a network G, where V is the set of nodes and E is the set of edges, we use DeepWalk, defined as f ( G ) , to obtain the representational vectors of nodes in V. The r v , v V is the self-representational vector of node v, which is calculated as follows:
{ r 0 , r 1 , , r v } = f ( G ) ;
Then, the nodes’ self-representational vectors are utilized to calculate the Euclidean distance between the nodes and their neighbors in embedding space, which serves as the aim of learning self-attention. By using weighted neighborhood aggregation, the local representational vectors of nodes are created from their neighbors’ self-representational vectors. Finally, the sequentially connected vectors of the self-representational vectors and the local representational vectors are fused into a single unified embedding space by Principle Component Analysis (PCA).

3.2. Weighted Neighborhood Aggregation

The process of neighborhood aggregation is frequently used to collect neighborhood information from adjacent nodes. The neighborhood information of node u is embedded as the local representational vector r u . The r u is calculated as follows:
r u = 1 | n ( u ) | v n ( u ) r v ,
where n ( u ) is the set of the neighbors of the node u, and r v is the self-representational vector of node v. DeeWaNA calculates the weights between nodes and incorporates the weights into the neighborhood aggregation to distinguish the influences of neighbor nodes on the central node. The local representational vector of node u, using weighted neighborhood aggregation, is calculated as follows:
r u = v n ( u ) α u v r v ;
where n ( u ) represents the set of neighbors of node u, α u v is the weight between node u and node v, and r v is the self-representational vector of node v. The next section presents three distinct weighting calculations.

3.3. Euclidean-Distance-Based Weighting Calculations

Self-attention is a widely used learning attention method to weight the various relationship between nodes and their neighbors [23]. To obtain the weights between nodes and their neighbors, we proposed three Euclidean-distance-based weighting calculations.

3.3.1. One-Head Self-Attention

The self-representational vectors of nodes and the vectors of their neighbors are used as inputs in the one-head self-attention model to determine the attention (weights), as shown in Figure 3. The attention function is described as follows:
A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d k ) V
where Q = I W Q , K = I W K , and V = I W V ; I R n d m o d e l is the matrix of representation, the learning matrices are W Q R d m o d e l d k , W K R d m o d e l d k , W V R d m o d e l d v , n = | I | is the size of inputs, and d m o d e l is the dimension of representation. The outputs are put into a full-connect Multi-Layer Perception (MLP). The MLP consists of three layers with ( n d m o d e l , n d m o d e l 2 , n 1 ) neurons. The estimated weights of node i by one-head self-attention are calculated as follows:
α i = s o f t m a x ( Q i K i T d k ) .
We utilize the distance in embedding space d between nodes as the target of the learning task. The output of the MLP is the estimated distance d . We aim to minimize the loss function as follows:
L o s s ( d , d ) = ( d d ) 2
for training the self-attention model.

3.3.2. Multi-Head Self-Attention

In contrast to the one-head self-attention model with queries (Q), keys (K), and values (V), multi-head self-attention attends to information from various representational subspaces, as depicted in Figure 4. Each subspace of representation is processed by the one-head self-attention function. The calculation of multi-head self-attention is described as follows:
M u l t i H e a d ( I ) θ = S p l i c e ( h e a d 1 , , h e a d h ) W o u t
where h e a d i = A t e e n t i o n ( I W i Q , I W i K , I W i V ) , θ = { W 1 h Q , W 1 h K , W 1 h V , W o u t } , W Q R d m o d e l d k , W K R d m o d e l d k , W V R d m o d e l d v , W o u t R h d v d m o d e l . Each subrepresentation is calculated by Equation (5). The value of the multi-head self-attention is calculated as follows:
α = 1 h i = 0 h α i
For the MLP structure, there are three layers with ( n d m o d e l , n d m o d e l 2 , n 1 ) neurons ( n = | I | ).

3.3.3. Non-Learning Weighting Calculation

Self-attention demands unacceptably long computation times and memory requisition on some large networks. To solve this issue, a straightforward weighting calculation is suggested and illustrated in Figure 5. The distance d u v between node u and node v is normalized into the (0, 1) range as follows:
d u v = 1 1 + e d u v
The weight between node v and node u is calculated by the following:
w u v = e d u v d u j e d u j , j { n ( u ) } ;
where n ( u ) is the set of neighbors of node u. The non-learning approach substitutes the self-attentional method to calculate the weights between nodes when a node has a large number of neighbors.

3.4. Connection and Fusion

DeeWaNA fuses the linked representational vectors into a single unified embedding space by PCA after splicing the local representational vectors after the self-representational vectors. This is because the simple connection of two different-view representational vectors may cause adverse effects on following network analysis tasks due to the representational vectors coming from different domains.

4. Experiments

We evaluated DeeWaNA on four real-world networks and compared it against four widely used unsupervised network representation learning algorithms, including (DeepWalk [6], LINE [7], Node2vec [9], and SDNE [8]). We set the dimension of representation in all the methods as 128 and used the two-step representation in the LINE. The restricted number of neighbors was set to 10 in all simplified versions. We set the number of heads in multi-head self-attention to 4. Other parameters were set as default in corresponding papers. All experiments were run on the platform of 2.8 GHz 4-core Intel i7, 16 GB 2133 MHz LPDDR3.

4.1. Datasets

We employed four real-world datasets (Blogcatalog [24], Protein-Protein Interactions (PPI) [25], Amherst, Mich [26]) to evaluate the performance of DeeWaNA, and the detailed statistics of these datasets are listed in Table 1.

4.2. Experimental Results

We utilized a one-vs-rest logistic regression model [6] to evaluate the quality of the learned representational vectors of all methods in node classification tasks. The training set was created by randomly selecting a proportion of the nodes, and the testing set was created using the remaining nodes. We repeated this process 10 times to calculate the average values on the Micro-F1.

4.2.1. Results on Multi-Label Classification Tasks

For all multi-label networks, DeeWaNA steadily performed better than other competitors within the rate of training data from 0.2 to 0.8. The experimental results of the multi-label classification tasks on Blogcatalog and PPI are displayed in Table 2 and Table 3.
On the Blogcatalog dataset, DeeWaNA with the non-learning weighting calculation produced the best performance (bold indicated) across all training proportions. One-head/multi-head self-attention was unable to produce the desired results due to memory limitations, as indicated by ‘-’. On the PPI dataset, DeeWaNA performed best at training proportions of (0.2, 0.8) when using the non-learning weighting calculation and at (0.4, 0.6) when using a combination of the multi-head self-attention and the non-learning calculation.

4.2.2. Results on Single-Label Classification Tasks

For all single-label networks, DeeWaNA consistently outperformed its competitors in the 0.2 to 0.8 range of training data proportions. The experimental results of single-label node classification tasks on Amherst and Mich are displayed in Table 4 and Table 5. When one-head self-attention and the non-learning calculation were coupled, DeeWaNA fared the best (bold indicated) in all training proportions on the Amherst dataset. On the Mich dataset, when utilizing the one-head self-attention and the non-learning calculation, DeeWaNA scored best at the training proportions of (0.4, 0.6, and 0.8) and at (0.2) when using the multi-head self-attention and the non-learning weighting calculation.
Our findings show that the various weighting calculations for DeeWaNA perform differently in the multi-label networks, while the combination of the one-head self-attention and the non-learning weighting calculation performs well in the majority of situations in the single-label networks. In single-label tasks, we advise utilizing one-head self-attention along with a non-learning weighting calculation, and in multi-label tasks, choose an appropriate weighting calculation.

4.2.3. Study on Effectiveness of Weighted Neighborhood Aggregation

To demonstrate the effectiveness of the weighted neighborhood aggregation, we conducted an experimental investigation applying weighted neighborhood aggregation to three widely used unsupervised network representation learning methods. LiNA is the technique that uses LINE instead of DeepWalk in DeeWaNA. NoNA is the method that substitutes the node2vec algorithm for DeepWalk in DeeWaNA. SeNA is the approach that replaces DeepWalk in DeeWaNA with the SDNE method. In this part, all methods utilized non-learning weighting calculation. The experimental results are displayed in Table 6.
The experimental results showed that, with the exception of the LINE algorithm on PPI and the node2vec method on Amherst at the 0.2 training proportion, the weighted neighborhood aggregation could improve the corresponding unsupervised learning algorithm in all cases. We speculate that the neighborhood aggregation amplifies the noise in the poor-quality self-representational vectors obtained by node2vec and SDNE in the low training proportion. In addition, comparing DeeWaNA with LiNA, NoNA, and SeNA, as described in Table 7, applying DeepWalk in DeeAWaNA was the best option for obtaining self-representational vectors for all cases except on the Mich dataset. Even though NoNA outperformed other algorithms on the Mich dataset in terms of the (0.2, 0.4, 0.8) training proportions, we still favor DeepWalk due to the node2vec’s good performance in NoNA depending on hyper-parameter tuning.

4.2.4. Study on Relationship Between Distance and Weights

To explore the relationship between distance and weights, we compared two kinds of non-learning weighting calculations in DeeWaNA on multi-/single-label classification tasks. DeeWaNA processed the weights being inversely proportional to distances. The calculation is as defined in Equation (9). DeeWaNA (F) processed the weights being proportional to distances. In DeeWaNA (F), the distance d u v between node u and node v is normalized into the (0, 1) range as follows:
d u v = 1 1 + e d u v
All DeeWaNA versions employed the non-learning weighting calculation in this part. The experimental results are displayed in Table 8. We found that the DeeWaNA (N) obtained the dominant position in all cases. This is also in line with our intuition that the farther the distance, the less the influence.

4.2.5. Study on Effectiveness of Features

We carried out an experimental investigation with two DeeWaNA variants on the Blogcatalog dataset to demonstrate the effectiveness of the local representational vectors. DeeWaNA (local) applied the local representational vectors as the final representations. DeeWaNA (cnt) regarded the unfused connection of self-representational vectors and local representational vectors as the final representations. All DeeWaNA versions employed the non-learning weighting calculation in this investigation. The experimental results are listed in Table 9. Comparing the ‘DeepWalk’ and the ‘DeeWaNA (local)’ results, the local-representational vectors have less expressive ability than the self-representational vectors. Comparing the ‘DeepWalk’ and the ‘DeeWaNA (cnt)’ results, the unfused connected representational vectors are more expressive than the self-representational vectors. Comparing the ‘DeeWaNA (cnt)’ and the ‘DeeWaNA (N)’ results, the performance of representation could be enhanced by fusing the linked representational vectors into a single unified embedding space using principal component analysis. The ablation experiment demonstrated the effectiveness of local representational vectors as assistants and the usefulness of the fusion process.

4.2.6. Study on Fusion

We compared DeeWaNA against two variants of DeeWaNA, which are DeeWaNA (AE) and DeeWaNA (ICA). In this part, all of DeeWaNA versions applied non-learning weighting calculation. The DeeWaNA (AE) consists of a five-layer (256-192-128-192-256) encoder–decoder neural network model as the fusion part. The DeeWaNA (ICA) used the Independent Component Analysis algorithm [27] as the fusion part. The experimental results are displayed in Table 10, and the best results are shown in bold. The results demonstrate that the PCA as the fusion process in DeeWaNA could obtain the best performance.

4.2.7. Study on Multi-Head Self-Attention

We investigated the 2/4/8 heads (MH-2/4/8) and the combined multi-head self-attention 2/4/8 heads with non-learning weighting calculations (MH-2L/4L/8L) to ascertain the number of heads in the multi-head self-attention models. The experimental results are displayed in Figure 6. Despite the fact that no particular number of heads predominated over others in all training proportions of the multi-head self-attention models, through the comparison of the stability and the worst case of prediction, we set the multi-head to four in this paper.

4.3. Limitations

Overall, the DeeWaNA could effectively obtain network representational vectors with good performance on the single-label and multi-label node classification tasks. The weighted neighborhood aggregation can easily be applied to improve some unsupervised network representation learning methods. But, there are some limitations in this work. First, for simplified calculations, only one-hop neighbors’ impacts were considered when DeeWaNA processed weighted neighborhood aggregation. The loss of ignoring multi-hop aggregation will be discussed in future works. Second, some high time-consuming or high-memory-required unsupervised network representation learning methods are not included in this work due to our hardware performance limitations. Third, more research is required before claiming weighted neighborhood aggregation can be applied generally, because this work lacks some SOTA unsupervised network representation learning methods that are particularly time- or memory-intensive. We think that despite many limitations, it is still a good attempt to give existing non-deep-structure-based unsupervised representation learning algorithms the ability to capture neighbor information.

5. Conclusions

In this study, we propose DeeWaNA, an unsupervised network representation learning framework that synergistically integrates DeepWalk with neighborhood-aware aggregation for enhanced node classification. The framework’s effectiveness stems from three key innovations: First, DeeWaNA enhances the conventional DeepWalk approach by incorporating neighborhood information awareness, thereby improving the discriminative power of learned node representations. Second, we introduce an adaptive weighted aggregation mechanism that dynamically quantifies heterogeneous neighbor influences during feature integration. Third, we develop a geometry-aware learning paradigm using Euclidean-distance-based self-attention to preserve spatial relationships in the embedding space.
Through comprehensive experimental evaluation on multiple real-world datasets across different domains, we demonstrate DeeWaNA’s superior performance against four unsupervised network representation learning baselines. The consistent outperformance across evaluation metrics reveals that DeeWaNA generates more semantically meaningful representations that significantly boost downstream node classification accuracy compared to existing approaches.

Author Contributions

Conceptualization/methodology/software/validation/writing—original draft preparation, X.X.; writing—review and editing, X.L. and J.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fundamental Research Funds for the Central Universities of the Ministry of Education of China (No. 2412024QS037).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Daokun, Z.; Jie, Y.; Xingquan, Z.; Chengqi, Z. Network Representation Learning: A Survey. IEEE Trans. Big Data 2017, 6, 3–28. [Google Scholar]
  2. Cannistraci, C.V.; Alanis-Lobato, G.; Ravasi, T. Minimum curvilinearity to enhance topological prediction of protein interactions by network embedding. Bioinformatics 2013, 29, i199–i209. [Google Scholar]
  3. Tu, C.; Wang, H.; Zeng, X.; Liu, Z.; Sun, M. Community-enhanced network representation learning for network analysis. arXiv 2016, arXiv:1611.06645. [Google Scholar]
  4. Xu, X.; Fu, Y.; Xiong, H.; Jin, B.; Li, X.; Hu, S.; Yin, M. Dr. right!: Embedding-based adaptively-weighted mixture multi-classification model for finding right doctors with healthcare experience data. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 647–656. [Google Scholar]
  5. Scarselli, F.; Gori, M.; Tsoi, A.C.; Hagenbuchner, M.; Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw. 2008, 20, 61–80. [Google Scholar] [CrossRef] [PubMed]
  6. Perozzi, B.; Al-Rfou, R.; Skiena, S. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 701–710. [Google Scholar]
  7. Tang, J.; Qu, M.; Wang, M.; Zhang, M.; Yan, J.; Mei, Q. Line: Large-scale information network embedding. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 1067–1077. [Google Scholar]
  8. Wang, D.; Cui, P.; Zhu, W. Structural deep network embedding. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1225–1234. [Google Scholar]
  9. Grover, A.; Leskovec, J. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
  10. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  11. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  12. Hamilton, W.; Ying, Z.; Leskovec, J. Inductive representation learning on large graphs. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 1024–1034. [Google Scholar]
  13. Hu, F.; Zhu, Y.; Wu, S.; Huang, W.; Wang, L.; Tan, T. GraphAIR: Graph representation learning with neighborhood aggregation and interaction. Pattern Recognit. 2021, 112, 107745. [Google Scholar] [CrossRef]
  14. Sun, Z.; Wang, C.; Hu, W.; Chen, M.; Qu, Y. Knowledge Graph Alignment Network with Gated Multi-Hop Neighborhood Aggregation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 222–229. [Google Scholar]
  15. Xie, Y.; Li, S.; Yang, C.; Wong, C.W.; Han, J. When Do GNNs Work: Understanding and Improving Neighborhood Aggregation. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence IJCAI-PRICAI-20, Yokohama, Japan, 11–17 July 2020. [Google Scholar]
  16. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM Trans. Graph. 2018, 38, 1–12. [Google Scholar] [CrossRef]
  17. Tipping, M.E.; Bishop, C.M. Probabilistic principal component analysis. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 1999, 61, 611–622. [Google Scholar] [CrossRef]
  18. Zhang, Y.; Tang, M. Consistency of random-walk based network embedding algorithms. arXiv 2021, arXiv:2101.07354. [Google Scholar]
  19. Guo, K.; Zhao, Z.; Yu, Z.; Guo, W.; Lin, R.; Tang, Y.; Wu, L. Network Embedding Based on Biased Random Walk for Community Detection in Attributed Networks. IEEE Trans. Comput. Soc. Syst. 2022, 10, 2279–2290. [Google Scholar] [CrossRef]
  20. Gallicchio, C.; Micheli, A. Graph Echo State Networks. In Proceedings of the 2010 International Joint Conference on Neural Networks (IJCNN), Barcelona, Spain, 18–23 July 2010; pp. 1–8. [Google Scholar]
  21. Li, Y.; Tarlow, D.; Brockschmidt, M.; Zemel, R. Gated Graph Sequence Neural Networks. arXiv 2017, arXiv:1511.05493. [Google Scholar]
  22. Shen, Y.; Li, H.; Li, D.; Zheng, J.; Wang, W. ANGraph: Attribute-interactive neighborhood-aggregative graph representation learning. Neural Comput. Appl. 2022, 34, 17937–17949. [Google Scholar] [CrossRef]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, N.A.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  24. Tang, L.; Liu, H. Relational learning via latent social dimensions. In Proceedings of the 15th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Paris, France, 28 June–1 July 2009; ACM: New York, NY, USA, 2009; pp. 817–826. [Google Scholar]
  25. Stark, C.; Breitkreutz, B.J.; Chatr-Aryamontri, A.; Boucher, L.; Oughtred, R.; Livstone, M.S.; Nixon, J.; Van Auken, K.; Wang, X.; Shi, X.; et al. The BioGRID interaction database: 2011 update. Nucleic Acids Res. 2010, 39, D698–D704. [Google Scholar] [CrossRef] [PubMed]
  26. Traud, A.L.; Mucha, P.J.; Porter, M.A. Social structure of facebook networks. Phys. A Stat. Mech. Its Appl. 2012, 391, 4165–4180. [Google Scholar] [CrossRef]
  27. Hyvärinen, A.; Oja, E. Independent component analysis: Algorithms and applications. Neural Netw. 2000, 13, 411–430. [Google Scholar] [CrossRef] [PubMed]
Figure 1. An example of node classification updated by different neighborhood aggregation operators: (a) Network G. (b) Neighborhood aggregation techniques can reduce difficulty of the classification task. (c) Poor neighborhood aggregation techniques will arrange neighbor nodes closely together in the embedding space. (d) Catastrophic aggregation techniques make tasks more difficult.
Figure 1. An example of node classification updated by different neighborhood aggregation operators: (a) Network G. (b) Neighborhood aggregation techniques can reduce difficulty of the classification task. (c) Poor neighborhood aggregation techniques will arrange neighbor nodes closely together in the embedding space. (d) Catastrophic aggregation techniques make tasks more difficult.
Entropy 27 00322 g001
Figure 2. Framework overview. First, DeepWalk is utilized to obtain self-representational vectors of nodes. Then, the weights between neighbors are calculated by the self-attention model on a Euclidean-distance-based learning task. The local representational vectors of nodes are calculated by the weighted neighborhood aggregation. Finally, the sequentially connected vectors are fused into a single unified embedding space.
Figure 2. Framework overview. First, DeepWalk is utilized to obtain self-representational vectors of nodes. Then, the weights between neighbors are calculated by the self-attention model on a Euclidean-distance-based learning task. The local representational vectors of nodes are calculated by the weighted neighborhood aggregation. Finally, the sequentially connected vectors are fused into a single unified embedding space.
Entropy 27 00322 g002
Figure 3. Framework of one-head self-attention: Each one of lines calculates the weights of one node’s neighbors. The top line is the process of calculation for the weights of neighbors of node v 0 , the middle line is for v 1 , and so on.
Figure 3. Framework of one-head self-attention: Each one of lines calculates the weights of one node’s neighbors. The top line is the process of calculation for the weights of neighbors of node v 0 , the middle line is for v 1 , and so on.
Entropy 27 00322 g003
Figure 4. The multi-head self-attention mechanism employs parallel computation units to independently derive attention weights for different nodes in the graph structure. Specifically, each attention head generates a distinct set of learnable projection parameters that computes the scaled dot-product attention between a target node and its neighbors. As illustrated in the computational graph, the uppermost attention head processes the edge weight relationships for node v 0 , and the intermediate head handles the neighborhood interactions for node v 1 , with this pattern extending to subsequent nodes. Crucially, all attention heads operate concurrently through parameter-independent transformation matrices, enabling the model to capture heterogeneous relational patterns in the feature space.
Figure 4. The multi-head self-attention mechanism employs parallel computation units to independently derive attention weights for different nodes in the graph structure. Specifically, each attention head generates a distinct set of learnable projection parameters that computes the scaled dot-product attention between a target node and its neighbors. As illustrated in the computational graph, the uppermost attention head processes the edge weight relationships for node v 0 , and the intermediate head handles the neighborhood interactions for node v 1 , with this pattern extending to subsequent nodes. Crucially, all attention heads operate concurrently through parameter-independent transformation matrices, enabling the model to capture heterogeneous relational patterns in the feature space.
Entropy 27 00322 g004
Figure 5. Non-learning weighting. From left to right: calculating the Euclidean distance between nodes and their neighbors; normalizing the distance into (0,1); through the Softmax function, obtaining the weights.
Figure 5. Non-learning weighting. From left to right: calculating the Euclidean distance between nodes and their neighbors; normalizing the distance into (0,1); through the Softmax function, obtaining the weights.
Entropy 27 00322 g005
Figure 6. The comparison of different heads of multi-head self-attention on the Amherst dataset.
Figure 6. The comparison of different heads of multi-head self-attention on the Amherst dataset.
Entropy 27 00322 g006
Table 1. Statistics of datasets.
Table 1. Statistics of datasets.
Dataset V E Multi# Labels
Blogcatalog10,312333,983Yes39
PPI389076,584Yes50
Amherst223590,954No6
Mich374581,901No10
Table 2. Experimental results of multi-label classification on Blogcatalog on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 2. Experimental results of multi-label classification on Blogcatalog on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
AlgorithmTraining Proportion
0.2 0.4 0.6 0.8
DeepWalk38.4940.6741.7842.51
LINE29.6633.7435.5436.76
node2vec37.9939.6940.7241.33
SDNE29.6230.4530.7530.95
DeeWaNA (N)39.4441.2742.2542.91
DeeWaNA (1H + N)39.3641.2342.1142.83
DeeWaNA (MH + N)39.3741.2542.1942.72
DeeWaNA (1H)----
DeeWaNA (MH)----
Table 3. Experimental results of multi-label classification on PPI on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 3. Experimental results of multi-label classification on PPI on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
AlgorithmTraining Proportion
0.2 0.4 0.6 0.8
DeepWalk18.3020.5822.0122.88
LINE12.2914.0415.4416.69
node2vec18.2320.3721.6622.72
SDNE16.8118.4719.3420.08
DeeWaNA (N)19.9922.4423.6724.55
DeeWaNA (1H + N)19.9122.3723.6824.54
DeeWaNA (MH + N)19.8522.5223.7024.47
DeeWaNA (1H)19.9622.3923.5724.54
DeeWaNA (MH)19.9722.3723.6824.47
Table 4. Experimental results of single-label classification on Amherst on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 4. Experimental results of single-label classification on Amherst on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
AlgorithmTraining Proportion
0.2 0.4 0.6 0.8
DeepWalk35.6137.2738.5839.60
LINE31.6034.3836.5237.63
node2vec36.3837.1837.9638.53
SDNE35.4236.8837.8938.88
DeeWaNA (N)37.5338.6639.3839.98
DeeWaNA (1H + N)37.6438.7439.5940.13
DeeWaNA (MH + N)37.4138.3939.1139.68
DeeWaNA (1H)37.5438.4539.0539.56
DeeWaNA (MH)37.5338.4339.1339.62
Table 5. Experimental results of single-label classification on Mich on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 5. Experimental results of single-label classification on Mich on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
AlgorithmTraining Proportion
0.2 0.4 0.6 0.8
DeepWalk39.7042.1643.6844.53
LINE35.3338.1139.8941.33
node2vec39.4042.0643.4444.20
SDNE36.4638.3839.6040.01
DeeWaNA (N)39.6242.1143.7044.64
DeeWaNA (1H + N)39.6042.2043.7244.78
DeeWaNA (MH + N)39.8042.1943.5344.70
DeeWaNA (1H)39.4941.9543.4744.48
DeeWaNA (MH)39.6542.0143.2944.69
Table 6. The experimental results of effectiveness of weighted neighborhood aggregation on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 6. The experimental results of effectiveness of weighted neighborhood aggregation on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
DatasetAlgorithmTraining Proportion
0.2 0.4 0.6 0.8
BlogcatalogLINE29.6633.7435.5436.76
LiNA34.5836.6037.7138.31
node2vec37.9939.6940.7241.33
NoNA38.8540.3641.1741.67
SDNE29.6230.4530.7530.95
SeNA29.2130.7031.5532.08
PPILINE12.2914.0415.4416.69
LiNA16.7018.3919.5220.18
node2vec18.2320.3721.6622.72
NoNA19.9422.0023.0723.78
SDNE16.8118.4719.3520.08
SeNA17.8620.2021.5122.31
AmherstLINE31.6034.3836.5237.63
LiNA36.3437.6138.6239.31
node2vec36.3837.1837.9638.53
NoNA36.3737.4838.1138.93
SDNE35.4236.8837.8938.88
SeNA37.3638.5439.1539.77
MichLINE35.3338.1139.8941.33
LiNA38.2140.4741.6442.57
node2vec39.4042.0643.4444.20
NoNA39.8442.3843.6444.80
SDNE36.4638.3839.6040.01
SeNA37.6139.3440.2540.78
Table 7. The experimental results of various algorithms for learning self-representational vectors on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 7. The experimental results of various algorithms for learning self-representational vectors on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
DatasetAlgorithmTraining Proportion
0.2 0.40.60.8
BlocatalogLiNA34.5836.6037.7138.31
NoNA38.8540.3641.1741.67
SeNA29.2130.7031.5532.08
De*NA (N)39.3641.2742.2542.91
PPILiNA16.7018.3919.5220.18
NoNA19.9422.0023.0723.78
SeNA17.8620.2021.5122.31
De*NA (N)19.9922.4423.6724.55
AmherstLiNA36.3437.6138.6239.31
NoNA36.3737.4838.1138.93
SeNA37.3638.5439.1539.77
De*NA (N)37.5338.6639.3839.98
MichLiNA38.2140.4741.6442.57
NoNA39.8442.3843.6444.80
SeNA37.6139.3440.2540.78
De*NA (N)39.6242.1143.7044.64
Table 8. The classification results of different distance–weight relationships on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 8. The classification results of different distance–weight relationships on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
DatasetAlgorithmTraining Proportion
0.20.40.60.8
BlogcatalogDeeWaNA (F)39.3641.2642.1442.74
DeeWaNA (N)39.4441.2742.2542.91
PPIDeeWaNA (F)19.8722.4223.6224.54
DeeWaNA (N)19.9922.4423.6724.55
AmherstDeeWaNA (F)37.4438.4839.2339.88
DeeWaNA (N)37.5338.6639.3839.98
MichDeeWaNA (F)39.6042.0843.6144.51
DeeWaNA (N)39.6242.1143.7044.64
Table 9. The experimental results of various combinations of representation on Blogcatalog on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 9. The experimental results of various combinations of representation on Blogcatalog on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
AlgorithmTraining Proportion
0.20.40.60.8
DeepWalk38.4940.6741.7842.51
DeeWaNA (local)36.1238.6039.6740.32
DeeWaNA (cnt)38.6840.8941.9642.67
DeeWaNA (N)39.4441.2742.2542.91
Table 10. The classification results of different fusion methods on Blogcatalog on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
Table 10. The classification results of different fusion methods on Blogcatalog on Micro-F1 (%). Bold indicates the optimal result among the comparative algorithms under the same parameters.
AlgorithmTraining Proportion
0.20.40.60.8
DeeWaNA (ICA)17.6418.6119.8121.12
DeeWaNA (AE)35.2137.6238.7339.45
DeeWaNA (N)39.3641.2742.2542.91
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, X.; Lu, X.; Wang, J. DeeWaNA: An Unsupervised Network Representation Learning Framework Integrating Deepwalk and Neighborhood Aggregation for Node Classification. Entropy 2025, 27, 322. https://doi.org/10.3390/e27030322

AMA Style

Xu X, Lu X, Wang J. DeeWaNA: An Unsupervised Network Representation Learning Framework Integrating Deepwalk and Neighborhood Aggregation for Node Classification. Entropy. 2025; 27(3):322. https://doi.org/10.3390/e27030322

Chicago/Turabian Style

Xu, Xin, Xinya Lu, and Jianan Wang. 2025. "DeeWaNA: An Unsupervised Network Representation Learning Framework Integrating Deepwalk and Neighborhood Aggregation for Node Classification" Entropy 27, no. 3: 322. https://doi.org/10.3390/e27030322

APA Style

Xu, X., Lu, X., & Wang, J. (2025). DeeWaNA: An Unsupervised Network Representation Learning Framework Integrating Deepwalk and Neighborhood Aggregation for Node Classification. Entropy, 27(3), 322. https://doi.org/10.3390/e27030322

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop