Graph Pooling with Node Proximity for Hierarchical Representation
Learning
- URL: http://arxiv.org/abs/2006.11118v1
- Date: Fri, 19 Jun 2020 13:09:44 GMT
- Title: Graph Pooling with Node Proximity for Hierarchical Representation
Learning
- Authors: Xing Gao, Wenrui Dai, Chenglin Li, Hongkai Xiong, Pascal Frossard
- Abstract summary: We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
- Score: 80.62181998314547
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph neural networks have attracted wide attentions to enable representation
learning of graph data in recent works. In complement to graph convolution
operators, graph pooling is crucial for extracting hierarchical representation
of graph data. However, most recent graph pooling methods still fail to
efficiently exploit the geometry of graph data. In this paper, we propose a
novel graph pooling strategy that leverages node proximity to improve the
hierarchical representation learning of graph data with their multi-hop
topology. Node proximity is obtained by harmonizing the kernel representation
of topology information and node features. Implicit structure-aware kernel
representation of topology information allows efficient graph pooling without
explicit eigendecomposition of the graph Laplacian. Similarities of node
signals are adaptively evaluated with the combination of the affine
transformation and kernel trick using the Gaussian RBF function. Experimental
results demonstrate that the proposed graph pooling strategy is able to achieve
state-of-the-art performance on a collection of public graph classification
benchmark datasets.
Related papers
- Graph Parsing Networks [64.5041886737007]
We propose an efficient graph parsing algorithm to infer the pooling structure, which then drives graph pooling.
The resulting Graph Parsing Network (GPN) adaptively learns personalized pooling structure for each individual graph.
arXiv Detail & Related papers (2024-02-22T09:08:36Z) - Saliency-Aware Regularized Graph Neural Network [39.82009838086267]
We propose the Saliency-Aware Regularized Graph Neural Network (SAR-GNN) for graph classification.
We first estimate the global node saliency by measuring the semantic similarity between the compact graph representation and node features.
Then the learned saliency distribution is leveraged to regularize the neighborhood aggregation of the backbone.
arXiv Detail & Related papers (2024-01-01T13:44:16Z) - Demystifying Graph Convolution with a Simple Concatenation [6.542119695695405]
We quantify the information overlap between graph topology, node features, and labels.
We show that graph concatenation is a simple but more flexible alternative to graph convolution.
arXiv Detail & Related papers (2022-07-18T16:39:33Z) - CGMN: A Contrastive Graph Matching Network for Self-Supervised Graph
Similarity Learning [65.1042892570989]
We propose a contrastive graph matching network (CGMN) for self-supervised graph similarity learning.
We employ two strategies, namely cross-view interaction and cross-graph interaction, for effective node representation learning.
We transform node representations into graph-level representations via pooling operations for graph similarity computation.
arXiv Detail & Related papers (2022-05-30T13:20:26Z) - Edge but not Least: Cross-View Graph Pooling [76.71497833616024]
This paper presents a cross-view graph pooling (Co-Pooling) method to better exploit crucial graph structure information.
Through cross-view interaction, edge-view pooling and node-view pooling seamlessly reinforce each other to learn more informative graph-level representations.
arXiv Detail & Related papers (2021-09-24T08:01:23Z) - Graph Pooling via Coarsened Graph Infomax [9.045707667111873]
We propose Coarsened GraphPool Infomaxing (CGI) to maximize the mutual information between the input and the coarsened graph of each pooling layer.
To achieve mutual information neural, we apply contrastive learning and propose a self-attention-based algorithm for learning positive and negative samples.
arXiv Detail & Related papers (2021-05-04T03:50:21Z) - Accurate Learning of Graph Representations with Graph Multiset Pooling [45.72542969364438]
We propose a Graph Multiset Transformer (GMT) that captures the interaction between nodes according to their structural dependencies.
Our experimental results show that GMT significantly outperforms state-of-the-art graph pooling methods on graph classification benchmarks.
arXiv Detail & Related papers (2021-02-23T07:45:58Z) - Second-Order Pooling for Graph Neural Networks [62.13156203025818]
We propose to use second-order pooling as graph pooling, which naturally solves the above challenges.
We show that direct use of second-order pooling with graph neural networks leads to practical problems.
We propose two novel global graph pooling methods based on second-order pooling; namely, bilinear mapping and attentional second-order pooling.
arXiv Detail & Related papers (2020-07-20T20:52:36Z) - Multilevel Graph Matching Networks for Deep Graph Similarity Learning [79.3213351477689]
We propose a multi-level graph matching network (MGMN) framework for computing the graph similarity between any pair of graph-structured objects.
To compensate for the lack of standard benchmark datasets, we have created and collected a set of datasets for both the graph-graph classification and graph-graph regression tasks.
Comprehensive experiments demonstrate that MGMN consistently outperforms state-of-the-art baseline models on both the graph-graph classification and graph-graph regression tasks.
arXiv Detail & Related papers (2020-07-08T19:48:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.