Extending the Design Space of Graph Neural Networks by Rethinking
Folklore Weisfeiler-Lehman
- URL: http://arxiv.org/abs/2306.03266v3
- Date: Sun, 14 Jan 2024 14:02:53 GMT
- Title: Extending the Design Space of Graph Neural Networks by Rethinking
Folklore Weisfeiler-Lehman
- Authors: Jiarui Feng, Lecheng Kong, Hao Liu, Dacheng Tao, Fuhai Li, Muhan
Zhang, Yixin Chen
- Abstract summary: Message passing neural networks (MPNNs) have emerged as the most popular framework of graph neural networks (GNNs) in recent years.
However, their expressive power is limited by the 1-dimensional Weisfeiler-Lehman (1-WL) test.
We propose an extension, $(k,t)$-FWL, which considers any equivariant set as neighbors instead of all nodes.
N$2$-GNN achieves record-breaking results on ZINC-Subset (0.059), outperforming previous SOTA results by 10.6%.
- Score: 66.23316415757456
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Message passing neural networks (MPNNs) have emerged as the most popular
framework of graph neural networks (GNNs) in recent years. However, their
expressive power is limited by the 1-dimensional Weisfeiler-Lehman (1-WL) test.
Some works are inspired by $k$-WL/FWL (Folklore WL) and design the
corresponding neural versions. Despite the high expressive power, there are
serious limitations in this line of research. In particular, (1) $k$-WL/FWL
requires at least $O(n^k)$ space complexity, which is impractical for large
graphs even when $k=3$; (2) The design space of $k$-WL/FWL is rigid, with the
only adjustable hyper-parameter being $k$. To tackle the first limitation, we
propose an extension, $(k,t)$-FWL. We theoretically prove that even if we fix
the space complexity to $O(n^k)$ (for any $k\geq 2$) in $(k,t)$-FWL, we can
construct an expressiveness hierarchy up to solving the graph isomorphism
problem. To tackle the second problem, we propose $k$-FWL+, which considers any
equivariant set as neighbors instead of all nodes, thereby greatly expanding
the design space of $k$-FWL. Combining these two modifications results in a
flexible and powerful framework $(k,t)$-FWL+. We demonstrate $(k,t)$-FWL+ can
implement most existing models with matching expressiveness. We then introduce
an instance of $(k,t)$-FWL+ called Neighborhood$^2$-FWL (N$^2$-FWL), which is
practically and theoretically sound. We prove that N$^2$-FWL is no less
powerful than 3-WL, and can encode many substructures while only requiring
$O(n^2)$ space. Finally, we design its neural version named N$^2$-GNN and
evaluate its performance on various tasks. N$^2$-GNN achieves record-breaking
results on ZINC-Subset (0.059), outperforming previous SOTA results by 10.6%.
Moreover, N$^2$-GNN achieves new SOTA results on the BREC dataset (71.8%) among
all existing high-expressive GNN methods.
Related papers
- Approximation Rates for Shallow ReLU$^k$ Neural Networks on Sobolev Spaces via the Radon Transform [4.096453902709292]
We consider the problem of how efficiently shallow neural networks with the ReLU$k$ activation function can approximate functions from Sobolev spaces.
We provide a simple proof of nearly optimal approximation rates in a variety of cases, including when $qleq p$, $pgeq 2$, and $s leq k + (d+1)/2$.
arXiv Detail & Related papers (2024-08-20T16:43:45Z) - Spatio-Spectral Graph Neural Networks [50.277959544420455]
We propose Spatio-Spectral Graph Networks (S$2$GNNs)
S$2$GNNs combine spatially and spectrally parametrized graph filters.
We show that S$2$GNNs vanquish over-squashing and yield strictly tighter approximation-theoretic error bounds than MPGNNs.
arXiv Detail & Related papers (2024-05-29T14:28:08Z) - Weisfeiler and Leman Go Loopy: A New Hierarchy for Graph Representational Learning [17.646846505225735]
We introduce a novel hierarchy of graph isomorphism tests and a corresponding GNN framework, $r$-$ell$MPNN, that can count cycles up to length $r + 2$.
Most notably, we show that $r$-$ell$WL can count homomorphisms of cactus graphs.
We empirically validate the expressive and counting power of the proposed $r$-$ell$MPNN on several synthetic datasets and present state-of-the-art predictive performance on various real-world datasets.
arXiv Detail & Related papers (2024-03-20T16:58:28Z) - Learning General Policies for Classical Planning Domains: Getting Beyond C$_2$ [15.574717738100727]
GNN-based approaches for learning general policies across planning domains are limited by the expressive power of $C$.
We introduce a parameterized version of relational GNNs, when $t$ is infinity, R-GNN[$t$] approximates $3$-GNNs using only quadratic space for embeddings.
For lower values of $t$, such as $t=1 and $t=2$, R-GNN[$t$] achieves a weaker approximation by exchanging fewer messages.
arXiv Detail & Related papers (2024-03-18T12:42:53Z) - Capacity Bounds for Hyperbolic Neural Network Representations of Latent
Tree Structures [8.28720658988688]
We study the representation capacity of deep hyperbolic neural networks (HNNs) with a ReLU activation function.
We establish the first proof that HNNs can embed any finite weighted tree into a hyperbolic space of dimensiond$ at least equal to $2$.
We find that the network complexity of HNN implementing the graph representation is independent of the representation fidelity/distortion.
arXiv Detail & Related papers (2023-08-18T02:24:32Z) - When Expressivity Meets Trainability: Fewer than $n$ Neurons Can Work [59.29606307518154]
We show that as long as the width $m geq 2n/d$ (where $d$ is the input dimension), its expressivity is strong, i.e., there exists at least one global minimizer with zero training loss.
We also consider a constrained optimization formulation where the feasible region is the nice local region, and prove that every KKT point is a nearly global minimizer.
arXiv Detail & Related papers (2022-10-21T14:41:26Z) - Neural Networks Efficiently Learn Low-Dimensional Representations with
SGD [22.703825902761405]
We show that SGD-trained ReLU NNs can learn a single-index target of the form $y=f(langleboldsymbolu,boldsymbolxrangle) + epsilon$ by recovering the principal direction.
We also provide compress guarantees for NNs using the approximate low-rank structure produced by SGD.
arXiv Detail & Related papers (2022-09-29T15:29:10Z) - Exponential Separation between Quantum and Classical Ordered Binary
Decision Diagrams, Reordering Method and Hierarchies [68.93512627479197]
We study quantum Ordered Binary Decision Diagrams($OBDD$) model.
We prove lower bounds and upper bounds for OBDD with arbitrary order of input variables.
We extend hierarchy for read$k$-times Ordered Binary Decision Diagrams ($k$-OBDD$) of width.
arXiv Detail & Related papers (2022-04-22T12:37:56Z) - Improving Robustness and Generality of NLP Models Using Disentangled
Representations [62.08794500431367]
Supervised neural networks first map an input $x$ to a single representation $z$, and then map $z$ to the output label $y$.
We present methods to improve robustness and generality of NLP models from the standpoint of disentangled representation learning.
We show that models trained with the proposed criteria provide better robustness and domain adaptation ability in a wide range of supervised learning tasks.
arXiv Detail & Related papers (2020-09-21T02:48:46Z) - Curse of Dimensionality on Randomized Smoothing for Certifiable
Robustness [151.67113334248464]
We show that extending the smoothing technique to defend against other attack models can be challenging.
We present experimental results on CIFAR to validate our theory.
arXiv Detail & Related papers (2020-02-08T22:02:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.