Next Article in Journal
Models for Quantum Measurement of Particles with Higher Spin
Previous Article in Journal
Towards Quantum Control with Advanced Quantum Computing: A Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Graph Multihead Attention Pooling with Self-Supervised Learning

College of Computer Science and Technology, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(12), 1745; https://doi.org/10.3390/e24121745
Submission received: 7 October 2022 / Revised: 16 November 2022 / Accepted: 25 November 2022 / Published: 29 November 2022

Abstract

:
Graph neural networks (GNNs), which work with graph-structured data, have attracted considerable attention and achieved promising performance on graph-related tasks. While the majority of existing GNN methods focus on the convolutional operation for encoding the node representations, the graph pooling operation, which maps the set of nodes into a coarsened graph, is crucial for graph-level tasks. We argue that a well-defined graph pooling operation should avoid the information loss of the local node features and global graph structure. In this paper, we propose a hierarchical graph pooling method based on the multihead attention mechanism, namely GMAPS, which compresses both node features and graph structure into the coarsened graph. Specifically, a multihead attention mechanism is adopted to arrange nodes into a coarsened graph based on their features and structural dependencies between nodes. In addition, to enhance the expressiveness of the cluster representations, a self-supervised mechanism is introduced to maximize the mutual information between the cluster representations and the global representation of the hierarchical graph. Our experimental results show that the proposed GMAPS obtains significant and consistent performance improvements compared with state-of-the-art baselines on six benchmarks from the biological and social domains of graph classification and reconstruction tasks.

1. Introduction

Over the past years, deep neural networks have achieved great success in various tasks such as computer vision [1], natural language processing [2], and video processing [3]. However, with the development of data science, the form of data is no longer limited to Euclidean-based data. Ubiquitous non-Euclidean graph data have entered the focus of researchers, such as social networks [4,5,6], biological networks [7,8], knowledge graphs [9], etc. Therefore, GNNs, which extend deep neural network techniques to graph data, have become a research hotspot of great interest [10,11,12]. In recent years, a myriad of GNN methods have been proposed by researchers, which can be categorized into two groups: spectral methods [13,14,15] and spatial methods [5,16]. For spectral methods, they have their theoretical basis in the field of graph signal processing and introduce filters as graph convolution for the purpose of noise reduction of graph signals. For spatial methods, graph convolution is defined by the connection of nodes in topological space, where feature information is directly transferred from one node to its neighbors. Due to their convincing performance and interpretability, GNNs have been widely applied to graph-related tasks, such as node classification [5,17,18], link prediction [19,20], and graph classification [13,21].
Although the majority of the existing work focuses on the design of convolutional operations, graph pooling, which maps nodes into a coarsened graph, is critical for capturing hierarchical structural information in graph classification tasks. However, unlike the pooling operations in traditional CNNs, graph pooling operations are challenging due to the fact that graphs usually contain diverse irregular topologies and no explicit spatial sequence of nodes [21]. A simple graph pooling operation is globally summing or averaging the representations of the nodes in the entire graph to generate a graph-level representation [16]. However, such a pooling operation simply treats all elements as equal, which ignores the differences in the relative importance of various node features. Furthermore, the inherently flat structure of the pooling operation restricts the ability to hierarchically represent the whole graph. To address these limitations, several differentiable hierarchical pooling methods [21,22,23,24] have been proposed to cluster or sample nodes in the graph layer by layer with a neural network architecture in an end-to-end scheme. However, all those methods have obvious drawbacks. In the node sampling approaches, each pooling layer unnecessarily discards a certain percentage of nodes, which causes unnecessary loss of node features and destroys the inherent substructure in the graph. In the node clustering approaches, the two main parts are clustering assignment and constructing a coarsened graph structure. However, these current heuristic-based linear clustering assignment approaches [21,25] are not adaptive to accurately learning the graph representation in a specific downstream task, so have room for improvement. We contend that unnecessary node discarding and ineffective coarsened graph creation are barely adequate because a well-defined graph pooling operation should prevent the loss of node properties and graph topology information.
In this work, to obtain an accurate representation of the graph, we needed design a graph pooling method that can compress and encode the node set into a coarsened graph, and construct the coarsened graph structure. To this end, following DiffPool [21], the graph pooling problem was considered as a node clustering assignment problem, where each cluster represents a node in the next-layer coarsened graph. Specifically, we divided the graph pooling problem into three main units: node clustering assignment, coarsened graph construction, and self-supervised mutual information module. First, to overcome the inability of simple neural network models to discriminate important nodes, a multihead attention mechanism [26] was adopted as the main component in the pooling operation. Second, the clustering assignment matrix could be approximated by the product of Q and K regarding the attention mechanism, which constructs the structure of the coarsened graph. In addition, to ensure the expressive ability of cluster representations, we introduced a self-supervised mechanism [27,28] to maximize the mutual information between cluster representations and the global representation of the hierarchical graph. By using these three units as basic modules, graph multihead attention pooling with self-supervised learning (GMAPS) was designed, which compresses the given node features and graph structure into a coarsened graph, encodes the clusters (nodes) in the coarsened graph, and constructs the coarsened graph structure. Finally, we experimentally validated the performance of GMAPS in a graph classification task on six datasets from the biological and social domains, where the proposed method outperformed seven state-of-the-art baselines. Our main contributions are summarized as follows:
  • We considered the graph pooling as a node clustering assignment problem. For accurate clustering assignment, a multihead attention mechanism based on GNN was introduced to sufficiently consider the connections between nodes in terms of features and the structure. Then, we derived Q and K in the attention model to generate the topology of the coarsened graph.
  • Self-supervised learning was adopted to maximize the mutual information between the cluster representations and hierarchically global representations, which further optimized the node representation in the coarsened graph.
  • Finally, the experimental results showed that the proposed GMAPS significantly outperformed baseline methods in the graph classification task on six publicly available datasets from the biological and social domains.
The rest of this paper is organized as follows: In Section 2, we briefly review the related studies about GNNs and graph pooling. Section 3 introduces some preliminaries and presents the details of our proposed model. In Section 4, experimental results on six benchmark datasets are shown and analyzed to highlight the benefits. Finally, we conclude the paper in Section 5.

2. Related Work

2.1. Graph Neural Networks

Existing GNNs models generally utilize the message-passing strategy [29] to encode node representations and have achieved promising performance on node classification [4], link prediction [30], and graph classification [31] tasks. Advances are often categorized into two branches: spectral approaches [13,14,15] and spatial approaches [5,16]. For spectral approaches, the graph convolutional operation is typically defined according to graph spectral theory. Ref. [32] suggested that a spectral filter can be approximated by a truncated expansion in terms of Chebyshev K-order polynomials of graph Laplacian. Later, ChebNet [15] used this K-localized convolution to define a convolutional neural network on graphs. The graph convolutional network (GCN) [5] further limits the K-localized convolution to K = 1 as the layer-wise convolutional operation and implements rich convolutional filter functions by stacking multiples of such layers [33]. To jointly consider the local and global consistency on graphs, deep graph CNN (DGCNN) [34] extended GCN by adding a convolutional operation with positive point-wise mutual information (PPMI) matrix. For spatial approaches, the convolutional operation typically directly aggregates the neighborhood information to the central node. Specifically, [4] proposed GraphSAGE, a general inductive framework that encodes representations by sampling a fixed-size set of local neighborhoods and aggregating their features by mean, LSTM, or pooling. GAT [17] incorporates the attention mechanism into the aggregation step and utilizes the self-attention strategy to assign different weights to aggregated neighborhoods. Inspired by GAT, many researchers have incorporated structured information to the transformer by developing the structure-aware self-attention mechanism [35,36,37]. However, to the bet of our knowledge, there is no previous study applying a transformer to graph pooling operations. More details about graph neural networks can be found in several comprehensive reviews [38,39,40].

2.2. Graph Pooling

Graph pooling is an essential unit in the hierarchical graph representation learning task, which captures node features and the hierarchical graph structure. Directly averaging or summing the node representations of the entire graph is the simplest pooling operation; however, it ignores the diverse weight of nodes and the hierarchical graph structure [16]. Existing graph pooling approaches can be broadly categorized into node sampling and node clustering approaches. For the node sampling approaches, they score the nodes by various mechanisms and then proportionally select the nodes with high scores as the nodes of the coarsened graph, while the nodes with low scores are discarded. gPool [22] employs a trainable projection vector to adaptively downsample a subset of nodes, yet ignores the graph structure information. To integrate node features and graph topology in the graph pooling layer, SAGPool [23] utilizes graph convolution to compute self-attention scores. Furthermore, ASAP [41] defines the local neighbors within a fixed receptive field as clusters, and then exploits an attention mechanism to compute the fitness scores based on local extremum information. However, this branching approach inevitably loses some important node feature information and graph topology information while discarding nodes. In contrast, node clustering approaches replace the node sampling as a fixed number of clusters with aggregated nodes. DiffPool [21] proposes a differentiable hierarchical clustering module by training an assignment matrix and a topology matrix of coarsened graphs. In order to achieve satisfactory performance, it appends an auxiliary link prediction objective and entropy regularization. Yet, this heuristic model cannot be adapted to match specific downstream tasks. MinCutPool [25] formulates a continuous relaxation of the normalized minCUT problem [42] based on the theory of spectral clustering and trains an adaptive clustering assignment by optimizing this objective function. However, it involves high computational complexity and many iterations.

3. Proposed Model

The key idea of the proposed GMAPS is that it enables the construction of graph pooling through a differentiable node assignment based on a multihead attention mechanism and a hierarchical objective based on maximizing mutual information. In this section, we present the overall GNN architecture and show the details of each module.

3.1. Preliminaries

A graph G is represented as ( V , E ) , where V = { v 0 , v 1 , v n } and E = { e 0 , e 1 , , e l } are the setd of nodes and edges, respectively; n and l denote the number of nodes and edges, respectively. Let A { 0 , 1 } n × n denote the adjacent matrix and X R n × f be the node feature matrix, where f is the dimension of node features. Given a set of labeled graphs G = { ( G 1 , y 1 ) , ( G 2 , y 2 ) , } where y i Y indicates the label corresponding to graph G i G , the goal of the graph classification task is to learn a mapping F : G Y that maps the set of graphs to the set of labels. In addition, the nodes (clusters) and the graph structure are changed after each pooling layer, so we further denote the adjacent matrix and hidden representation matrix of graph G i fed into the kth layer as A i k R n i k × n i k and H i k R n i k × d , respectively, where d is the dimension of any hidden representations in neural networks, and n i k means the number of nodes in G i at layer k.

3.1.1. Graph Neural Networks

GNNs learn node representations through various aggregation schemes, which are generally described as the following message-passing architecture [29]:
H k + 1 = γ k ( H k , φ k ( H k , A k ) ) ,
where γ ( · ) denotes a differentiable update function, and φ ( · ) denotes a differentiable, permutation-invariant aggregation function, e.g., sum, mean, or max. H k is the node representation fed into the kth layer. The aggregation function φ ( · ) aggregates the representations of neighboring nodes into an aggregated representation. Then, the update function γ ( · ) concatenates (or sums) the current node representations with the aggregated representations as the updated node representations. Particularly, the input node representation H 1 is initialized using the node features on the graph, e.g., H 1 = X .
There are many possible implementations for the GNN architecture, such as GCN [5], GraphSAGE [4], and GAT [17]. In this study, we implemented the pooling operations on top of the GCN architecture, due to its classical and efficient nature. For the ( k + 1 ) th layer in GCN, the message-passing architecture can be formalized as follows:
H k + 1 = ReLU ( D ˜ 1 2 A ˜ D ˜ 1 2 H k W k ) ,
where ReLU ( · ) is the nonlinear activation function, A ˜ = A + I is the adjacent matrix with self-loops, D ˜ i i = j A ˜ i j is the diagonal degree matrix, and W k R d × d is a trainable weight matrix.

3.1.2. Graph Pooling

GNNs are inherently flat, as they only propagate information across the edge of a graph. The goal of graph pooling is to define a differentiable end-to-end operation to generate the structure and the representation of a coarsened graph for hierarchical GNN models. Following Ying’s strategy [21], the pooling problem can be considered as using the output of the GNN module to learn how to cluster or sample nodes to generate a coarsened graph so that we can use this coarsened graph as the input to the next layer. We denote the learned clustering assignment matrix at kth pooling layer as S k R n k × n k + 1 , and the node (cluster) representations H k + 1 and adjacent matrix A k + 1 of the coarsened graph of the k + 1 layer can be generated as follows:
H k + 1 = S k H k R n k + 1 × d ,
A k + 1 = S k A k S k R n k + 1 × n k + 1 ,
where · denotes the matrix transpose. Equation (3) uses the clustering assignment S k to softly cluster the kth layer nodes H k to generate the coarsened clusters (nodes) H k + 1 at k + 1 layer. Similarly, Equation (4) generates the adjacent matrix A k + 1 of the coarsened graph of the k + 1 layer to represent the connection strength between clusters (nodes).

3.2. Overall Neural Network Architecture

Figure 1 illustrates the framework of the proposed graph multihead attention pooling with self-supervised learning (GMAPS), which is implemented interleaved with graph convolutional operations to build a stacking GNNs architecture. GMAPS can be divided into three major components as node clustering assignment, coarsened graph construction, and self-supervised mutual information module. The whole GNNs architecture generally alternately stacks multiple graph convolutional layers and graph pooling layers in a hierarchical fashion. Then, a readout function is used to aggregate and concatenate the representations of each convolutional layer to generate the graph-level representation. Finally, the graph-level representations are fed into a multilayer perceptron (MLP) for the graph classification task. In the following, we detail each of the three major components of the proposed pooling method.

3.3. Node Clustering Assignment

Existing methods [21,25] generally utilize GNNs and MLPs to learn node clustering assignment; however, the former ignores the internal connections between nodes, and the latter ignores nonlinear relationships, which limit the performance of the clustering. In this study, to overcome these limitations, the attention mechanism was adopted to learn node clustering assignments.
Let n c denote the number of next-layer clusters (nodes), and we have queries Q R n c × d , keys K R n × d and values V R n × d as the input. The scaled dot-product attention [26] function can be formalized as follows:
Attention Q , K , V = softmax ( Q K d ) V .
In this paper, queries Q and keys K denote clusters and nodes, respectively. The dot-product operation Q K calculates the correlation between them, followed by a normalized function softmax ( · ) . Finally, based on the normalized weights, a weighted summation of the values V of the nodes is performed to obtain the representations of clusters.
Furthermore, instead of performing a single attention function with d-dimensional queries, keys, and values, we can linearly separate the queries, keys, and values with h different learned projections function into d / h dimensions. Multihead attention allows the model to jointly attend to information from different representation subspaces. The output of the multihead attention function is computed as follows:
MultiHead Q , K , V = Concat head 1 , , head h W O , head i = Attention ( Q W i Q , K W i K , V W i V ) ,
where the learned linear projections are parameter matrices W i Q R d × d , W i K R d × d , and W i V R d × d , and the concatenation projection is parameter matrix W O R d × d .
There is a challenge in directly implementing the above multihead attention mechanism into the graph pooling operation: while the multihead attention mechanism considers the internal connections between nodes, the linear projection of keys and values still inhibits the improvement in the attention. For this reason, inspired by DiffPool [21], we employed GNN models to generate keys and values in the multihead attention mechanism, such that it jointly considers the graph structure and the internal connections between nodes. Given the input node features H k R n k × d and coarsened adjacency matrix A k R n k × n k at layer k, the inputs of keys and values in the multihead attention can be generated as follows:
Keys = GNN k , key ( H k , A k ) , Values = GNN k , value ( H k , A k ) .
Note that the GNN models for keys and values can be employed based on specific tasks, and GCN [5], GraphSAGE [4], and GAT [17] all can adapt to the model. In contrast to the two GNNs in DiffPool that directly and linearly generate node embeddings and clustering assignments, separately, the two GNNs called key and value were designed to consider graph structure information in the graph multiheaded attention mechanism, which then enable the attention mechanism to learn the optimal node clustering assignments.
Based on the above ingredients, graph multihead attention (GMA) can be formally expressed as follows:
GMA Q , H , A = Concat head 1 , , head h W O , head i = Attention ( Q W i Q , GNN k , key ( H k , A k ) , GNN k , value ( H k , A k ) ) ,
where Q R n k + 1 × d is a parameterized seed matrix that clusters the n k nodes into n k + 1 clusters.
In addition to GMA, following the transformer model architecture, each of the proposed pooling layer contains a fully connected feed-forward network (FFN), which is separately and identically applied to each row, followed by layer normalization. Thus, the overall architecture of graph multihead attention pooling (GMAP) can be formally described as follows:
GMAP Z = LayerNorm Z + FNN Z , w h e r e Z = LayerNorm ( Q + GMA ( Q , H , A ) ) .
Note that the output of GMAP is the embeddings of the clusters in the coarsened graph, which serve as the node embeddings of the next graph convolutional layer. After compressing the nodes of the original graph into clusters, which are nodes in the coarsened graph, we need to reduce the adjacency matrix of the original graph to another refined adjacency matrix of the coarsened graph.

3.4. Coarsened Graph Construction

While a coarsened graph can be constructed by compressed clusters, the connections between nodes in the coarsened graph still need to be generated by the node clustering assignment matrix. We first consider the formation for node clustering assignment matrix by the attention function as in Equation (5). In this attention function, Q , K , and V denote the queries of clusters, the keys of nodes, and the values of nodes, respectively. The attention function in Equation (5) can be decomposed as follows:
Attention Q , K , V = softmax Q K d × V ,
where the first term softmax Q K d is the scaled dot product of cluster queries with all keys of nodes followed by a softmax function, namely the weights on the corresponding values of nodes. Specifically, the dot product Q K denotes the correlation between the clusters and nodes. The function softmax ( · ) normalizes the correlation to the correlation weights. Thus, the first term is essentially a generated soft cluster assignment matrix S :
S = softmax Q K d .
Note that compared with the cluster assignment directly generated by GNNs in DiffPool [21], our proposed model can be considered as explicitly learning the data-dependent n c cluster centroids by specific learnable queries Q , and then using the queries Q and keys K to generate the cluster assignment.
For each pair of nodes in the coarsened graph, the edges are constructed by considering the edges between the corresponding clusters in the original graph. Therefore, we can formally construct the adjacency matrix of the coarsened graph as follows:
A k + 1 = S k A k S k .

3.5. Self-Supervised Mutual Information Module

Our pooling method, GMAP, can achieve hierarchical representation of the original graph. To discriminate the cluster representations in the hierarchical graph, the self-supervised learning mechanism is utilized to maximize the mutual information between the cluster representation and the hierarchically global representation.
To generate the global representation of the hierarchical graph, we leveraged a readout function to summarize the node representation of the hierarchical coarsened graph into a fixed size graph level representation as follows:
r k = Readout H k = σ ( 1 n k p = 1 n k H k ( p , : ) ) ,
where σ ( · ) is a sigmoid function, and n k denotes the number of nodes in the kth layer graph.
As a proxy for maximizing the mutual information between the cluster representations and hierarchically coarsened graph, a discriminator function is employed to measure the probability scores that represent the clusters contained within the hierarchical coarsened graph. We formally show the discriminator function as follows:
D ( H i , : k , r k ) = σ ( H i , : k W D r k ) ,
where σ ( · ) is a sigmoid function, H i , : k denotes the embedding of node i, and W D R d × d is the linear parametric matrix.
For the self-supervised mutual information objective, we followed the intuitions in [28,43] and employed a noise-contrastive-type objective with a standard binary cross-entropy (BCE) loss between positive samples H k from the coarsened graph and the negative samples H ˜ k from another graph in the same batch. Therefore, the objective can be defined as follows:
L M I k = 1 n p o s + n n e g ( i = 1 n p o s E p o s l o g D H i , : k , r k + j = 1 n n e g E n e g l o g D H ˜ j , : k , r k ) ,
where n p o s and n n e g denote the number of positive and negative examples, respectively. Note that, the mutual information objective of GMAPS is complementary to the graph classification objective, which enables the proposed pooling layer, GMAP, to focus on both hierarchical and global structural properties.

3.6. Computational Complexity

The space complexity of each graph pooling layer is O ( n n c ) , as it depends on the soft cluster assignment matrix S R n × n c . The computational complexity of node clustering assignment is O ( n 2 d + n d 2 ) , which is dominated by a multihead attention operation. The computational complexity of coarsened graph construction is O ( n 2 n c + n n c 2 ) . Because the adjacency matrix is usually sparse, the computational complexity is reduced to O ( | E | n c + n n c 2 ) , where | E | is the number of nonzero edges in the adjacency matrix.

4. Experiments

To evaluate the performance of the proposed pooling model, GMAPS, GMAPS was compared with a collection of state-of-the-art GNN-based models on six benchmarks in terms of graph classification tasks. Furthermore, detailed ablation and parametric analyses were conducted to characterize the proposed pooling model.

4.1. Datasets

We adopted six public benchmark graph classification datasets among the TUDatasets [44]. The datasets are publicly available at https://chrsmrrs.github.io/datasets/docs/datasets/ (accessed on 22 June 2022). Statistics and properties are summarized in Table 1 with a detailed description as follows: D&D and PROTEINS are two protein graph datasets, MUTAG is a small molecule graph dataset, and the remaining ones are the social network datasets. Note that there are no node features in three social network datasets, and we encoded the node degrees into one-hot vectors as node features, which explicitly concerned the structural information.

4.2. Baselines

GMAPS was compared with state-of-the-art GNN-based models that could be categorized into two groups: global graph neural networks and graph pooling models.
Global graph neural networks include three representative models: GCN [5], GraphSAGE [4], and GAT [17], which learn node-level representations. Therefore, to achieve graph-level representations, we employed a readout function to summarize the learned node representations to a fixed-size graph representation.
For the graph pooling models, we deployed the same hierarchical GNN architecture, replacing only the pooling methods in it. To evaluate the performance of the proposed pooling model, gPool [22], ASAP [41], SAGPool [23], and DiffPool [21] were adopted as comparison methods. Specifically, gPool and SAGPool both feed the coarsened graph by a heuristic measure strategy to select a subset of nodes, the former by a trainable projection vector and the latter by attention between node features and graph topology. Furthermore, ASAP first considers the subgraphs within the fixed receptive field as clusters and then uses the self-attention among the local structure to measure the fitness scores for coarsened graphs. DiffPool learns a clustering assignment matrix with extra GNNs models to cluster nodes and to generate the adjacency matrix of the coarsened graph.
Finally, to further analyze the performance of the proposed pooling models, we introduced two variants: GMAP (Equation (9)) condenses the nodes to clusters with graph multihead attention mechanism, and GMAPS additionally employs self-supervised mutual information maximization to ensure the uniformity of the cluster representations in a hierarchical coarsened graph and the global representation of a coarsened graph.

4.3. Implementation Details

Following many previous studies [21,45], for all baselines, 10-fold cross-validation was used to randomly split each dataset into training, validation, and test sets in a ratio of 80%/10%/10%. Then, we performed the randomly splitting process 10 times with 10 different seeds, and we report the average accuracy with standard deviation. For baseline methods, we implemented the source code released by PyTorch Geometric (PyG) [46], and the hyperparameters followed the default setting. To achieve a fair comparison, all baselines were implemented on top of the same GNN architecture. In our study, the hierarchical GNN architecture consisted of three convolutional layers, two pooling layers, a readout module, and an MLP module. Two pooling layers were interspersed between convolutional layers, separately, and the jumping knowledge strategy [47] was employed to concentrate the hierarchical representations generated by each convolutional layer. The MLP model consisted of three fully connected layers and a softmax classifier. For three global GNN models (GCN, GraphSAGE, and GAT), the GNN architecture consisted of three corresponding convolutional layers, followed by a global mean readout function. The architectures are illustrated in Figure 2.
For all datasets, the learning rate was set to 5 × 10 4 , and weight decay was set to 1 × 10 4 . Because D&D has a large number of average nodes, its hidden size and batch size were discreetly set to 32 and 10, respectively. For the other datasets, the hidden size was set to 128, and batch size ws set to 128. The pooling ratio was set to 25% for all baselines. For the MLP, the hidden size of the three fully connected layers were separately set to 128 and 56, the number of graph classification labels, and the dropout ratio was set to 50%. Furthermore, early stopping was implemented with a patience parameter of 50, where the training process stopped if the performance on the validation set did not improve beyond 50 epochs.

4.4. Graph Classification

The results of graph classification are reported in Table 2. As we can see, GMAPS consistently outperformed all baselines on all datasets, and GMAP outperformed other baseline methods or achieved comparable performance to the other baseline methods. For instance, GMAPS achieved 9.5% and 13.8% improvement on D&D and MUTAG datasets over GCN without hierarchical pooling, respectively. These results demonstrate that the proposed hierarchical pooling strategies are powerful and versatile.
For global GNN baseline methods without hierarchical pooling, such as GCN, GraphSAGE, and GAT, their overall performance was weaker than that of the hierarchical graph pooling approaches, especially on D&D and MUTAG datasets. For such bioinformatic networks, the structure of the subgraphs therein often represents different properties of the molecule and is crucial for learning the graph representations. Global GNN baseline methods only globally summarize the node representations and ignore the graph structural information. Thus, they struggle to achieve satisfactory performance.
Among the hierarchical graph pooling models, DiffPool achieved superior results in most cases, while the other baseline methods had their own performance on various datasets. We argue that the major reason is that their respective heuristic node selection strategies are only applicable to specific scenarios and are not versatile. Specifically, DiffPool utilizes an auxiliary link prediction objective during training to cluster the nearby nodes. In contrast, our proposed model considers both the graph structure and the internal connections between nodes; the satisfactory performance on different datasets further validates the generality of GMAPS. Note that the performance on D&D, PROTEINS, and MUTAG showed an impressive improvement over that of the global baseline method, while the performance on IMDB-B, IMDB-M, and COLLAB was similar to that of the baseline method. This may be due to the fact that the latter three social network datasets lack raw features and encode node degrees as features.
Finally, among our proposed models, GMAPS, which employs self-supervised mutual information maximization, consistently outperformed GMAP, which indicated the mutual information maximum module is facilitative for graph pooling in the graph classification task.

4.5. Graph Reconstruction

To quantity the amount of information about the graph retained by different pooling methods, we trained an autoencoder to reconstruct the input node features from the pooled representations. The learning objective to minimize the mean squared error between the original feature X and the reconstructed features X r e c was defined as | | X X r e c | | 2 . To upscale the coarsened graph back to its original graph, for node clustering methods (DiffPool and GMAP), we transposed the pooling operation as X r e c = S X p o o l . For the top-K method (gPool), we used the unpooling operation proposed by [22]. We plot the visualization results in Figure 3 with auxiliary numerical measures (two-dimensional mean squared error). A good pooling method should recover the original graph as much as possible. It is obvious that gPool failed to recover the entire graph information because the top-K pooling operation dropped portions of the graph. The noisy result of DiffPool indicated a partial loss of information in the coarsened graph, while GMAP produced an almost perfect reconstruction result, which demonstrated its power in retaining meaningful information.

4.6. Ablation Studies and Visualization

As mentioned above, our proposed pooling models can be integrated into various GNN architectures. To explore their performance with different graph convolutional strategies, we implemented GraphSAGE and GAT, except for the default GCN convolutional layer. The experiments were conducted on the D&D, PROTEINS, and MUTAG datasets, and the results are presented in Table 3. As shown in Table 3, the graph classification performance depends not only on the pooling strategy but also on the choice of the graph convolutional strategy. In particular, GMAPS outperformed GMAP for any graph convolutional configuration on any dataset, which further confirmed the optimization of the self-supervised mutual information maximization module for the pooling layer in the graph classification task.
To further explore the clustering assignment process of the pooling model, a visualization of the clustering assignment matrix S is presented in Figure 4. We randomly selected a graph in the MUTAG dataset to feed into the DiffPool and GMAPS models to obtain the assignment matrix in the first pooling layer. As shown in Figure 4a, DiffPool tended to generate a smooth coarsened graph through a dense clustering assignment matrix. In contrast, in Figure 4b, GMAPS has a relatively sparse cluster assignment matrix, which allows for distinct node features and graph structure in the coarsened graphs. The visualization once again demonstrated the satisfactory performance of GMAPS in capturing node features and graph topology.

4.7. Parameters Analysis

We further investigated the effect of two hyperparameters (dimension d and pooling ratio r) on the pooling model, GMAPS, by considering the PROTEIN, MUTAG, and IMDB-BINARY datasets with various settings. Figure 5 shows that for the PROTEIN and IMDB-BINARY datasets, too large and too small dimensions diminished the accuracy, and 128 was the optimal choice. However, for the MUTAG dataset, due to its relatively simple graph number and structure, a lower dimension catered to its node representations, and the accuracy decreased instead as the dimension rose. The hyperparameter analysis of the pooling rate r is presented in Figure 6. It is shown that for the three experimental datasets, the accuracies improved with increasing pooling rate, which means that the pooling rate cannot be set too small; otherwise, most of the graph structure information will be lost, and, thus, the accuracy in the graph classification tasks will be hampered.

5. Conclusions

In this study, we found that existing graph pooling methods either cause the unnecessary loss of node features or do not adaptively learn an accurate graph representation in a specific downstream task. To address these limitations, we designed a novel graph pooling operation called GMAPS, which can compress nodes into a coarsened graph in a soft clustering manner. It also utilizes self-supervised mutual information maximization to ensure consistency between the cluster representations and the hierarchical coarsened graph representation. To validate the performance of the proposed pooling operation, we conducted experiments on six publicly available datasets and compared the results with those of seven state-of-the-art baseline methods. The results of graph classification and graph reconstruction tasks showed the excellent performance of the proposed model. Considering that the proposed pooling approach is generally applicable to various graph-learning tasks, which are growing more crucial, we are certain that it will have a significant practical impact. In the future, we will attempt to employ local subgraphs or anonymous random walks to capture the coarsened graph’s structure for more accurate pooling operations because improved local subgraph property extraction can improve the coarsened graph’s accuracy and interpretability. The key challenge of this effort is the capture and representation of task-related local subgraphs’ properties, which will be the focus of our future study. In addition, applying self-supervised learning without contrastive pairs to graph classification tasks is another aim for future research.

Author Contributions

Conceptualization, Y.W. (Yu Wang) and Y.W. (Yang Wu); methodology, Y.W. (Yu Wang); software, Y.W. (Yu Wang); validation, Y.W. (Yu Wang), Y.W. (Yang Wu) and L.H.; formal analysis, W.G.; investigation, W.G.; resources, Y.W. (Yu Wang); data curation, Y.W. (Yu Wang); writing—original draft preparation, Y.W. (Yu Wang); writing—review and editing, W.G.; supervision, W.G.; project administration, L.H.; funding acquisition, L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by the National Natural Science Foundation of China under grant no. 61701190; the Science Foundation of Jilin Province of China under grant no. 2020122209JC, Youth Sci-Tech Innovation Leader and Team Project of Jilin Province of China under grant no. 20170519017JH; Key Technology Innovation Cooperation Project of Government and University for the whole Industry Demonstration under grant no. SXGJSF2017-4; Key scientific and technological R&D Plan of Jilin Province of China under grant no. 20180201103GX; and the Project of Jilin Province Development and Reform Commission no. 2019FGWTZC001.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in an open access repository at [44].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  2. Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation by Jointly Learning to Align and Translate. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  3. Karpathy, A.; Toderici, G.; Shetty, S.; Leung, T.; Sukthankar, R.; Li, F. Large-Scale Video Classification with Convolutional Neural Networks. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014, Columbus, OH, USA, 23–28 June 2014; pp. 1725–1732. [Google Scholar] [CrossRef] [Green Version]
  4. Hamilton, W.L.; Ying, Z.; Leskovec, J. Inductive Representation Learning on Large Graphs. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 1024–1034. [Google Scholar]
  5. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 24–26 April 2017. [Google Scholar]
  6. Shang, Y. Consensus formation in networks with neighbor-dependent synergy and observer effect. Commun. Nonlinear Sci. Numer. Simul. 2021, 95, 105632. [Google Scholar] [CrossRef]
  7. Min, S.; Gao, Z.; Peng, J.; Wang, L.; Qin, K.; Fang, B. STGSN—A Spatial–Temporal Graph Neural Network framework for time-evolving social networks. Knowl.-Based Syst. 2021, 214, 106746. [Google Scholar] [CrossRef]
  8. Duvenaud, D.; Maclaurin, D.; Aguilera-Iparraguirre, J.; Gómez-Bombarelli, R.; Hirzel, T.; Aspuru-Guzik, A.; Adams, R.P. Convolutional Networks on Graphs for Learning Molecular Fingerprints. In Proceedings of the Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; pp. 2224–2232. [Google Scholar]
  9. Jiang, D.; Wang, R.; Yang, J.; Xue, L. Kernel multi-attention neural network for knowledge graph embedding. Knowl.-Based Syst. 2021, 227, 107188. [Google Scholar] [CrossRef]
  10. Niepert, M.; Ahmed, M.; Kutzkov, K. Learning Convolutional Neural Networks for Graphs. In Proceedings of the 33nd International Conference on Machine Learning, ICML 2016, New York, NY, USA, 19–24 June 2016; pp. 2014–2023. [Google Scholar]
  11. Wu, Y.; Song, Y.; Huang, H.; Ye, F.; Xie, X.; Jin, H. Enhancing Graph Neural Networks via auxiliary training for semi-supervised node classification. Knowl.-Based Syst. 2021, 220, 106884. [Google Scholar] [CrossRef]
  12. Yin, R.; Li, K.; Zhang, G.; Lu, J. A deeper graph neural network for recommender systems. Knowl.-Based Syst. 2019, 185, 105020. [Google Scholar] [CrossRef]
  13. Bruna, J.; Zaremba, W.; Szlam, A.; LeCun, Y. Spectral Networks and Locally Connected Networks on Graphs. In Proceedings of the 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  14. Henaff, M.; Bruna, J.; LeCun, Y. Deep Convolutional Networks on Graph-Structured Data. CoRR 2015. Available online: http://xxx.lanl.gov/abs/1506.05163 (accessed on 25 March 2022).
  15. Defferrard, M.; Bresson, X.; Vandergheynst, P. Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering. In Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, Barcelona, Spain, 5–10 December 2016; pp. 3837–3845. [Google Scholar]
  16. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How Powerful are Graph Neural Networks? In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  17. Velickovic, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  18. Wang, Y.; Hu, L.; Gao, W. Learning flexible network representation via anonymous walks. Knowl. Based Syst. 2021, 222, 107021. [Google Scholar] [CrossRef]
  19. Schlichtkrull, M.S.; Kipf, T.N.; Bloem, P.; van den Berg, R.; Titov, I.; Welling, M. Modeling Relational Data with Graph Convolutional Networks. In Proceedings of the Semantic Web—15th International Conference, ESWC 2018, Heraklion, Greece, 3–7 June 2018; Volume 10843, pp. 593–607. [Google Scholar] [CrossRef] [Green Version]
  20. Vashishth, S.; Sanyal, S.; Nitin, V.; Talukdar, P.P. Composition-based Multi-Relational Graph Convolutional Networks. In Proceedings of the 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  21. Ying, Z.; You, J.; Morris, C.; Ren, X.; Hamilton, W.L.; Leskovec, J. Hierarchical Graph Representation Learning with Differentiable Pooling. In Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018, Montréal, QC, Canada, 3–8 December 2018; pp. 4805–4815. [Google Scholar]
  22. Gao, H.; Ji, S. Graph U-Nets. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 2083–2092. [Google Scholar]
  23. Lee, J.; Lee, I.; Kang, J. Self-Attention Graph Pooling. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 3734–3743. [Google Scholar]
  24. Gao, X.; Xiong, H.; Frossard, P. iPool—Information-based Pooling in Hierarchical Graph Neural Networks. CoRR 2019. Available online: http://xxx.lanl.gov/abs/1907.00832 (accessed on 23 May 2022). [CrossRef] [PubMed]
  25. Bianchi, F.M.; Grattarola, D.; Alippi, C. Spectral Clustering with Graph Neural Networks for Graph Pooling. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, Virtual Event, 13–18 July 2020; Volume 119, pp. 874–883. [Google Scholar]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  27. Jin, W.; Derr, T.; Liu, H.; Wang, Y.; Wang, S.; Liu, Z.; Tang, J. Self-supervised Learning on Graphs: Deep Insights and New Direction. CoRR 2020. Available online: http://xxx.lanl.gov/abs/2006.10141 (accessed on 23 May 2022).
  28. Belghazi, M.I.; Baratin, A.; Rajeswar, S.; Ozair, S.; Bengio, Y.; Hjelm, R.D.; Courville, A.C. Mutual Information Neural Estimation. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholmsmässan, Stockholm, Sweden, 10–15 July 2018; pp. 530–539. [Google Scholar]
  29. Gilmer, J.; Schoenholz, S.S.; Riley, P.F.; Vinyals, O.; Dahl, G.E. Neural Message Passing for Quantum Chemistry. In Proceedings of the 34th International Conference on Machine Learning, ICML 2017, Sydney, NSW, Australia, 6–11 August 2017; pp. 1263–1272. [Google Scholar]
  30. Schütt, K.; Kindermans, P.; Felix, H.E.S.; Chmiela, S.; Tkatchenko, A.; Müller, K. SchNet: A continuous-filter convolutional neural network for modeling quantum interactions. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 991–1001. [Google Scholar]
  31. Zhang, Z.; Bu, J.; Ester, M.; Zhang, J.; Yao, C.; Yu, Z.; Wang, C. Hierarchical Graph Pooling with Structure Learning. CoRR 2019. Available online: http://xxx.lanl.gov/abs/1911.05954 (accessed on 23 May 2022).
  32. Hammond, D.K.; Vandergheynst, P.; Gribonval, R. Wavelets on Graphs via Spectral Graph Theory. CoRR 2009. Available online: http://xxx.lanl.gov/abs/0912.3848 (accessed on 23 May 2022). [CrossRef] [Green Version]
  33. Shang, Y. Generalized k-core percolation on correlated and uncorrelated multiplex networks. Phys. Rev. E 2020, 101, 042306. [Google Scholar] [CrossRef] [PubMed]
  34. Zhang, M.; Cui, Z.; Neumann, M.; Chen, Y. An End-to-End Deep Learning Architecture for Graph Classification. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, LA, USA, 2–7 February 2018; McIlraith, S.A., Weinberger, K.Q., Eds.; AAAI Press: Palo Alto, CA, USA, 2018; pp. 4438–4445. [Google Scholar]
  35. Wu, L.; Chen, Y.; Shen, K.; Guo, X.; Gao, H.; Li, S.; Pei, J.; Long, B. Graph Neural Networks for Natural Language Processing: A Survey. CoRR 2021. Available online: http://xxx.lanl.gov/abs/2106.06090 (accessed on 24 May 2022).
  36. Yao, S.; Wang, T.; Wan, X. Heterogeneous Graph Transformer for Graph-to-Sequence Learning. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, 5–10 July 2020; pp. 7145–7154. [Google Scholar] [CrossRef]
  37. Cai, D.; Lam, W. Graph Transformer for Graph-to-Sequence Learning. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; AAAI Press: Palo Alto, CA, USA, 2020; pp. 7464–7471. [Google Scholar]
  38. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Yu, P.S. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Networks Learn. Syst. 2021, 32, 4–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Sato, R. A Survey on The Expressive Power of Graph Neural Networks. CoRR 2020. Available online: http://xxx.lanl.gov/abs/2003.04078 (accessed on 25 May 2022).
  40. Errica, F.; Podda, M.; Bacciu, D.; Micheli, A. A Fair Comparison of Graph Neural Networks for Graph Classification. In Proceedings of the 8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  41. Ranjan, E.; Sanyal, S.; Talukdar, P.P. ASAP: Adaptive Structure Aware Pooling for Learning Hierarchical Graph Representations. In Proceedings of the Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, the Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; AAAI Press: Palo Alto, CA, USA, 2020; pp. 5470–5477. [Google Scholar]
  42. Dhillon, I.S.; Guan, Y.; Kulis, B. Kernel k-means: Spectral clustering and normalized cuts. In Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; Kim, W., Kohavi, R., Gehrke, J., DuMouchel, W., Eds.; ACM: New York, NY, USA, 2004; pp. 551–556. [Google Scholar] [CrossRef]
  43. Velickovic, P.; Fedus, W.; Hamilton, W.L.; Liò, P.; Bengio, Y.; Hjelm, R.D. Deep Graph Infomax. In Proceedings of the 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  44. Morris, C.; Kriege, N.M.; Bause, F.; Kersting, K.; Mutzel, P.; Neumann, M. TUDataset: A collection of benchmark datasets for learning with graphs. CoRR 2020. Available online: http://xxx.lanl.gov/abs/2007.08663 (accessed on 23 May 2022).
  45. Ma, Y.; Wang, S.; Aggarwal, C.C.; Tang, J. Graph Convolutional Networks with EigenPooling. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2019, Anchorage, AK, USA, 4–8 August 2019; Teredesai, A., Kumar, V., Li, Y., Rosales, R., Terzi, E., Karypis, G., Eds.; ACM: New York, NY, USA, 2019; pp. 723–731. [Google Scholar] [CrossRef]
  46. Fey, M.; Lenssen, J.E. Fast Graph Representation Learning with PyTorch Geometric. CoRR 2019. Available online: http://xxx.lanl.gov/abs/1903.02428 (accessed on 21 May 2022).
  47. Xu, K.; Li, C.; Tian, Y.; Sonobe, T.; Kawarabayashi, K.; Jegelka, S. Representation Learning on Graphs with Jumping Knowledge Networks. In Proceedings of the 35th International Conference on Machine Learning, ICML 2018, Stockholm, Sweden, 10–15 July 2018; Volume 80, pp. 5449–5458. [Google Scholar]
Figure 1. The overall neural network architecture of proposed GMAPS.
Figure 1. The overall neural network architecture of proposed GMAPS.
Entropy 24 01745 g001
Figure 2. Illustration of architectures of (a) hierarchical GNN and (b) global GNN (GCN, GraphSAGE, and GAT).
Figure 2. Illustration of architectures of (a) hierarchical GNN and (b) global GNN (GCN, GraphSAGE, and GAT).
Entropy 24 01745 g002
Figure 3. Reconstruction results of the ring graph: (a) original, (b) gPool (error = 1), (c) DiffPool (error = 10 2 ), and (d) GMAP (error = 10 6 ).
Figure 3. Reconstruction results of the ring graph: (a) original, (b) gPool (error = 1), (c) DiffPool (error = 10 2 ), and (d) GMAP (error = 10 6 ).
Entropy 24 01745 g003
Figure 4. Visualization of the assignment matrices of DiffPool and GMAPS. Rows represent clusters in the coarsened graph, while columns are the nodes in the original graph: (a) DiffPool and (b) GMAPS.
Figure 4. Visualization of the assignment matrices of DiffPool and GMAPS. Rows represent clusters in the coarsened graph, while columns are the nodes in the original graph: (a) DiffPool and (b) GMAPS.
Entropy 24 01745 g004
Figure 5. The parameter analysis of dimension d.
Figure 5. The parameter analysis of dimension d.
Entropy 24 01745 g005
Figure 6. The parameter analysis of pooling ratio r.
Figure 6. The parameter analysis of pooling ratio r.
Entropy 24 01745 g006
Table 1. Statistics of the datasets.
Table 1. Statistics of the datasets.
Dataset | G | Avg . | V | Avg . | E | | Y |
D&D1178284.32715.662
PROTEINS111339.0672.822
MUTAG18817.9319.792
IMDB-Binary100019.7796.532
IMDB-Multi150013.0065.943
COLLAB500074.492457.783
Table 2. Results of graph classification in terms of average accuracy ± standard deviation.
Table 2. Results of graph classification in terms of average accuracy ± standard deviation.
MethodD&DPROTEINSMUTAGIMDB-BIMDB-MCOLLAB
GCN-mean71.96 ± 4.7873.24 ± 3.6270.55 ± 10.2572.51 ± 3.9151.13 ± 3.1680.51 ± 1.39
SAGE-mean72.13 ± 2.6271.44 ± 3.8868.88 ± 15.1572.21 ± 2.6349.81 ± 3.9379.66 ± 1.49
GAT-mean71.36 ± 4.3172.61 ± 4.9169.44 ± 14.9572.59 ± 2.8350.46 ± 4.3179.36 ± 1.67
gPool74.87 ± 3.8272.52 ± 3.8172.77 ± 9.7671.91 ± 3.3650.26 ± 3.3379.11 ± 2.15
ASAP72.82 ± 3.1771.17 ± 4.8180.55 ± 10.0172.51 ± 4.2950.21 ± 5.2377.52 ± 2.38
SAGPool70.59 ± 3.1170.36 ± 3.8178.33 ± 8.7670.41 ± 5.9351.26 ± 3.5978.89 ± 1.63
DiffPool77.35 ± 3.4172.97 ± 5.7677.77 ± 9.2970.59 ± 4.6550.66 ± 3.6779.43 ± 1.42
GMAP77.61 ± 3.4174.23 ± 4.3281.67 ± 11.1272.09 ± 4.0151.33 ± 4.6780.72 ± 1.38
GMAPS78.81 ± 4.0574.41 ± 3.5883.33 ± 8.9572.61 ± 4.3851.67 ± 5.4180.97 ± 1.41
Table 3. Results of GMAP and GMAPS with various convolutional strategies.
Table 3. Results of GMAP and GMAPS with various convolutional strategies.
MethodD&DPROTEINSMUTAG
GMAP-GCN77.61 ± 3.4174.23 ± 4.3281.67 ± 11.12
GMAP-GraphSAGE76.92 ± 3.4872.07 ± 4.3680.42 ± 10.58
GMAP-GAT76.79 ± 4.1572.97 ± 4.9180.11 ± 11.62
GMAPS-GCN78.81 ± 4.0574.41 ± 3.5883.33 ± 8.95
GMAPS-GraphSAGE78.63 ± 3.1172.97 ± 3.7581.96 ± 9.72
GMAPS-GAT78.37 ± 3.1973.87 ± 4.8881.25 ± 10.7
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, Y.; Hu, L.; Wu, Y.; Gao, W. Graph Multihead Attention Pooling with Self-Supervised Learning. Entropy 2022, 24, 1745. https://doi.org/10.3390/e24121745

AMA Style

Wang Y, Hu L, Wu Y, Gao W. Graph Multihead Attention Pooling with Self-Supervised Learning. Entropy. 2022; 24(12):1745. https://doi.org/10.3390/e24121745

Chicago/Turabian Style

Wang, Yu, Liang Hu, Yang Wu, and Wanfu Gao. 2022. "Graph Multihead Attention Pooling with Self-Supervised Learning" Entropy 24, no. 12: 1745. https://doi.org/10.3390/e24121745

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop