On the Bottleneck of Graph Neural Networks and its Practical
Implications
- URL: http://arxiv.org/abs/2006.05205v4
- Date: Tue, 9 Mar 2021 11:53:06 GMT
- Title: On the Bottleneck of Graph Neural Networks and its Practical
Implications
- Authors: Uri Alon, Eran Yahav
- Abstract summary: We show that graph neural networks (GNNs) are susceptible to a bottleneck when aggregating messages across a long path.
This bottleneck causes the over-squashing of exponentially growing information into fixed-size vectors.
GNNs fail to propagate messages originating from distant nodes and perform poorly when the prediction task depends on long-range interaction.
- Score: 22.704284264177108
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Since the proposal of the graph neural network (GNN) by Gori et al. (2005)
and Scarselli et al. (2008), one of the major problems in training GNNs was
their struggle to propagate information between distant nodes in the graph. We
propose a new explanation for this problem: GNNs are susceptible to a
bottleneck when aggregating messages across a long path. This bottleneck causes
the over-squashing of exponentially growing information into fixed-size
vectors. As a result, GNNs fail to propagate messages originating from distant
nodes and perform poorly when the prediction task depends on long-range
interaction. In this paper, we highlight the inherent problem of over-squashing
in GNNs: we demonstrate that the bottleneck hinders popular GNNs from fitting
long-range signals in the training data; we further show that GNNs that absorb
incoming edges equally, such as GCN and GIN, are more susceptible to
over-squashing than GAT and GGNN; finally, we show that prior work, which
extensively tuned GNN models of long-range problems, suffers from
over-squashing, and that breaking the bottleneck improves their
state-of-the-art results without any tuning or additional weights. Our code is
available at https://github.com/tech-srl/bottleneck/ .
Related papers
- Spatio-Spectral Graph Neural Networks [50.277959544420455]
We propose Spatio-Spectral Graph Networks (S$2$GNNs)
S$2$GNNs combine spatially and spectrally parametrized graph filters.
We show that S$2$GNNs vanquish over-squashing and yield strictly tighter approximation-theoretic error bounds than MPGNNs.
arXiv Detail & Related papers (2024-05-29T14:28:08Z) - LazyGNN: Large-Scale Graph Neural Networks via Lazy Propagation [51.552170474958736]
We propose to capture long-distance dependency in graphs by shallower models instead of deeper models, which leads to a much more efficient model, LazyGNN, for graph representation learning.
LazyGNN is compatible with existing scalable approaches (such as sampling methods) for further accelerations through the development of mini-batch LazyGNN.
Comprehensive experiments demonstrate its superior prediction performance and scalability on large-scale benchmarks.
arXiv Detail & Related papers (2023-02-03T02:33:07Z) - Distributed Graph Neural Network Training: A Survey [51.77035975191926]
Graph neural networks (GNNs) are a type of deep learning models that are trained on graphs and have been successfully applied in various domains.
Despite the effectiveness of GNNs, it is still challenging for GNNs to efficiently scale to large graphs.
As a remedy, distributed computing becomes a promising solution of training large-scale GNNs.
arXiv Detail & Related papers (2022-11-01T01:57:00Z) - Training Graph Neural Networks on Growing Stochastic Graphs [114.75710379125412]
Graph Neural Networks (GNNs) rely on graph convolutions to exploit meaningful patterns in networked data.
We propose to learn GNNs on very large graphs by leveraging the limit object of a sequence of growing graphs, the graphon.
arXiv Detail & Related papers (2022-10-27T16:00:45Z) - Addressing Over-Smoothing in Graph Neural Networks via Deep Supervision [13.180922099929765]
Deep graph neural networks (GNNs) suffer from over-smoothing when the number of layers increases.
We propose DSGNNs enhanced with deep supervision where representations learned at all layers are used for training.
We show that DSGNNs are resilient to over-smoothing and can outperform competitive benchmarks on node and graph property prediction problems.
arXiv Detail & Related papers (2022-02-25T06:05:55Z) - Two-level Graph Neural Network [15.014364222532347]
We propose a novel GNN framework, referred to as the Two-level GNN (TL-GNN)
This merges subgraph-level information with node-level information.
Experiments show that TL-GNN outperforms existing GNNs and achieves state-of-the-art performance.
arXiv Detail & Related papers (2022-01-03T02:15:20Z) - Graph-less Neural Networks: Teaching Old MLPs New Tricks via
Distillation [34.676755383361005]
Graph-less Neural Networks (GLNNs) have no inference graph dependency.
We show that GLNNs with competitive performance infer faster than GNNs by 146X-273X and faster than other acceleration methods by 14X-27X.
A comprehensive analysis of GLNN shows when and why GLNN can achieve competitive results to Gs and suggests GLNN as a handy choice for latency-constrained applications.
arXiv Detail & Related papers (2021-10-17T05:16:58Z) - Overcoming Catastrophic Forgetting in Graph Neural Networks [50.900153089330175]
Catastrophic forgetting refers to the tendency that a neural network "forgets" the previous learned knowledge upon learning new tasks.
We propose a novel scheme dedicated to overcoming this problem and hence strengthen continual learning in graph neural networks (GNNs)
At the heart of our approach is a generic module, termed as topology-aware weight preserving(TWP)
arXiv Detail & Related papers (2020-12-10T22:30:25Z) - AutoGraph: Automated Graph Neural Network [45.94642721490744]
We propose a method to automate the deep Graph Neural Networks (GNNs) design.
In our proposed method, we add a new type of skip connection to the GNNs search space to encourage feature reuse.
We also allow our evolutionary algorithm to increase the layers of GNNs during the evolution to generate deeper networks.
arXiv Detail & Related papers (2020-11-23T09:04:17Z) - Gated Graph Recurrent Neural Networks [176.3960927323358]
We introduce Graph Recurrent Neural Networks (GRNNs) as a general learning framework for graph processes.
To address the problem of vanishing gradients, we put forward GRNNs with three different gating mechanisms: time, node and edge gates.
The numerical results also show that GRNNs outperform GNNs and RNNs, highlighting the importance of taking both the temporal and graph structures of a graph process into account.
arXiv Detail & Related papers (2020-02-03T22:35:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.