Discovering the Representation Bottleneck of Graph Neural Networks from
Multi-order Interactions
- URL: http://arxiv.org/abs/2205.07266v2
- Date: Tue, 17 May 2022 06:53:46 GMT
- Title: Discovering the Representation Bottleneck of Graph Neural Networks from
Multi-order Interactions
- Authors: Fang Wu, Siyuan Li, Lirong Wu, Stan Z. Li, Dragomir Radev, Qiang Zhang
- Abstract summary: Graph neural networks (GNNs) rely on the message passing paradigm to propagate node features and build interactions.
Recent works point out that different graph learning tasks require different ranges of interactions between nodes.
We study two common graph construction methods in scientific domains, i.e., emphK-nearest neighbor (KNN) graphs and emphfully-connected (FC) graphs.
- Score: 51.597480162777074
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most graph neural networks (GNNs) rely on the message passing paradigm to
propagate node features and build interactions. Recent works point out that
different graph learning tasks require different ranges of interactions between
nodes. To investigate its underlying mechanism, we explore the capacity of GNNs
to capture pairwise interactions between nodes under contexts with different
complexities, especially for their graph-level and node-level applications in
scientific domains like biochemistry and physics. When formulating pairwise
interactions, we study two common graph construction methods in scientific
domains, i.e., \emph{K-nearest neighbor} (KNN) graphs and
\emph{fully-connected} (FC) graphs. Furthermore, we demonstrate that the
inductive bias introduced by KNN-graphs and FC-graphs hinders GNNs to learn the
most informative order of interactions. {Such a phenomenon is broadly shared by
several GNNs for different graph learning tasks and forbids GNNs to achieve the
global minimum loss, so we name it a \emph{representation bottleneck}.} To
overcome that, we propose a novel graph rewiring approach based on the pairwise
interaction strengths to dynamically adjust the reception fields of each node.
Extensive experiments in molecular property prediction and dynamic system
forecast prove the superiority of our method over state-of-the-art GNN
baselines. More importantly, this paper provides a reasonable explanation of
why subgraphs play an important role in the determination of graph properties.
Related papers
- Unveiling Global Interactive Patterns across Graphs: Towards Interpretable Graph Neural Networks [31.29616732552006]
Graph Neural Networks (GNNs) have emerged as a prominent framework for graph mining.
This paper proposes a novel intrinsically interpretable scheme for graph classification.
Global Interactive Pattern (GIP) learning introduces learnable global interactive patterns to explicitly interpret decisions.
arXiv Detail & Related papers (2024-07-02T06:31:13Z) - Harnessing Collective Structure Knowledge in Data Augmentation for Graph Neural Networks [25.12261412297796]
Graph neural networks (GNNs) have achieved state-of-the-art performance in graph representation learning.
We propose a novel approach, namely collective structure knowledge-augmented graph neural network (CoS-GNN)
arXiv Detail & Related papers (2024-05-17T08:50:00Z) - DGNN: Decoupled Graph Neural Networks with Structural Consistency
between Attribute and Graph Embedding Representations [62.04558318166396]
Graph neural networks (GNNs) demonstrate a robust capability for representation learning on graphs with complex structures.
A novel GNNs framework, dubbed Decoupled Graph Neural Networks (DGNN), is introduced to obtain a more comprehensive embedding representation of nodes.
Experimental results conducted on several graph benchmark datasets verify DGNN's superiority in node classification task.
arXiv Detail & Related papers (2024-01-28T06:43:13Z) - DEGREE: Decomposition Based Explanation For Graph Neural Networks [55.38873296761104]
We propose DEGREE to provide a faithful explanation for GNN predictions.
By decomposing the information generation and aggregation mechanism of GNNs, DEGREE allows tracking the contributions of specific components of the input graph to the final prediction.
We also design a subgraph level interpretation algorithm to reveal complex interactions between graph nodes that are overlooked by previous methods.
arXiv Detail & Related papers (2023-05-22T10:29:52Z) - Relation Embedding based Graph Neural Networks for Handling
Heterogeneous Graph [58.99478502486377]
We propose a simple yet efficient framework to make the homogeneous GNNs have adequate ability to handle heterogeneous graphs.
Specifically, we propose Relation Embedding based Graph Neural Networks (RE-GNNs), which employ only one parameter per relation to embed the importance of edge type relations and self-loop connections.
arXiv Detail & Related papers (2022-09-23T05:24:18Z) - Automatic Relation-aware Graph Network Proliferation [182.30735195376792]
We propose Automatic Relation-aware Graph Network Proliferation (ARGNP) for efficiently searching GNNs.
These operations can extract hierarchical node/relational information and provide anisotropic guidance for message passing on a graph.
Experiments on six datasets for four graph learning tasks demonstrate that GNNs produced by our method are superior to the current state-of-the-art hand-crafted and search-based GNNs.
arXiv Detail & Related papers (2022-05-31T10:38:04Z) - Graph Neural Networks for Graphs with Heterophily: A Survey [98.45621222357397]
We provide a comprehensive review of graph neural networks (GNNs) for heterophilic graphs.
Specifically, we propose a systematic taxonomy that essentially governs existing heterophilic GNN models.
We discuss the correlation between graph heterophily and various graph research domains, aiming to facilitate the development of more effective GNNs.
arXiv Detail & Related papers (2022-02-14T23:07:47Z) - Graph Neural Networks with Feature and Structure Aware Random Walk [7.143879014059894]
We show that in typical heterphilous graphs, the edges may be directed, and whether to treat the edges as is or simply make them undirected greatly affects the performance of the GNN models.
We develop a model that adaptively learns the directionality of the graph, and exploits the underlying long-distance correlations between nodes.
arXiv Detail & Related papers (2021-11-19T08:54:21Z) - Multi-grained Semantics-aware Graph Neural Networks [13.720544777078642]
Graph Neural Networks (GNNs) are powerful techniques in representation learning for graphs.
This work proposes a unified model, AdamGNN, to interactively learn node and graph representations.
Experiments on 14 real-world graph datasets show that AdamGNN can significantly outperform 17 competing models on both node- and graph-wise tasks.
arXiv Detail & Related papers (2020-10-01T07:52:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.