Empowering GNNs via Edge-Aware Weisfeiler-Leman Algorithm
- URL: http://arxiv.org/abs/2206.02059v3
- Date: Tue, 23 Jan 2024 08:06:35 GMT
- Title: Empowering GNNs via Edge-Aware Weisfeiler-Leman Algorithm
- Authors: Meng Liu, Haiyang Yu, Shuiwang Ji
- Abstract summary: Message passing graph neural networks (GNNs) are known to have their expressiveness upper-bounded by 1-dimensional Weisfeiler-Leman (1-WL) algorithm.
We propose a general and provably powerful GNN framework that preserves the scalability of the message passing scheme.
- Score: 79.68451803954751
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Message passing graph neural networks (GNNs) are known to have their
expressiveness upper-bounded by 1-dimensional Weisfeiler-Leman (1-WL)
algorithm. To achieve more powerful GNNs, existing attempts either require ad
hoc features, or involve operations that incur high time and space
complexities. In this work, we propose a general and provably powerful GNN
framework that preserves the scalability of the message passing scheme. In
particular, we first propose to empower 1-WL for graph isomorphism test by
considering edges among neighbors, giving rise to NC-1-WL. The expressiveness
of NC-1-WL is shown to be strictly above 1-WL and below 3-WL theoretically.
Further, we propose the NC-GNN framework as a differentiable neural version of
NC-1-WL. Our simple implementation of NC-GNN is provably as powerful as
NC-1-WL. Experiments demonstrate that our NC-GNN performs effectively and
efficiently on various benchmarks.
Related papers
- Rethinking GNN Expressive Power Research in the Machine Learning Community: Limitations, Issues, and Corrections [21.683245760896313]
We argue that using well-defined computational models is a reasonable approach to characterizing and exploring GNNs' expressive power.
We show that the WL test is not locally computable and is misaligned with the message-passing GNNs.
We highlight several open problems regarding GNN expressive power for further exploration.
arXiv Detail & Related papers (2024-10-02T08:01:50Z) - Union Subgraph Neural Networks [7.922920885565194]
We empower Graph Neural Networks (GNNs) by injecting neighbor-connectivity information extracted from a new type of substructure.
By infusing the encoded neighbor connectivities, we propose a novel model, namely Union Subgraph Neural Network (UnionSNN)
Experiments on 18 benchmarks of both graph-level and node-level tasks demonstrate that UnionSNN outperforms state-of-the-art baseline models.
arXiv Detail & Related papers (2023-05-25T05:52:43Z) - Graph Neural Networks are Inherently Good Generalizers: Insights by
Bridging GNNs and MLPs [71.93227401463199]
This paper pinpoints the major source of GNNs' performance gain to their intrinsic capability, by introducing an intermediate model class dubbed as P(ropagational)MLP.
We observe that PMLPs consistently perform on par with (or even exceed) their GNN counterparts, while being much more efficient in training.
arXiv Detail & Related papers (2022-12-18T08:17:32Z) - Two-Dimensional Weisfeiler-Lehman Graph Neural Networks for Link
Prediction [61.01337335214126]
Link prediction is one important application of graph neural networks (GNNs)
Most existing GNNs for link prediction are based on one-dimensional Weisfeiler-Lehman (1-WL) test.
We study a completely different approach which can directly obtain node pair (link) representations based on textittwo-dimensional Weisfeiler-Lehman (2-WL) tests.
arXiv Detail & Related papers (2022-06-20T04:50:38Z) - Expressiveness and Approximation Properties of Graph Neural Networks [6.1323142294300625]
We provide an elegant way to obtain bounds on the separation power of GNNs in terms of the Weisfeiler-Leman (WL) tests.
We use tensor language to define Higher-Order Message-Passing Neural Networks (or k-MPNNs), a natural extension of MPNNs.
Our approach provides a toolbox with which GNN architecture designers can analyze the separation power of their GNNs.
arXiv Detail & Related papers (2022-04-10T11:33:04Z) - From Stars to Subgraphs: Uplifting Any GNN with Local Structure
Awareness [23.279464786779787]
We introduce a general framework to uplift any MPNN to be more expressive.
Our framework is strictly more powerful than 1&2-WL, and is not less powerful than 3-WL.
Our method sets new state-of-the-art performance by large margins for several well-known graph ML tasks.
arXiv Detail & Related papers (2021-10-07T19:08:08Z) - Identity-aware Graph Neural Networks [63.6952975763946]
We develop a class of message passing Graph Neural Networks (ID-GNNs) with greater expressive power than the 1-WL test.
ID-GNN extends existing GNN architectures by inductively considering nodes' identities during message passing.
We show that transforming existing GNNs to ID-GNNs yields on average 40% accuracy improvement on challenging node, edge, and graph property prediction tasks.
arXiv Detail & Related papers (2021-01-25T18:59:01Z) - The Surprising Power of Graph Neural Networks with Random Node
Initialization [54.4101931234922]
Graph neural networks (GNNs) are effective models for representation learning on relational data.
Standard GNNs are limited in their expressive power, as they cannot distinguish beyond the capability of the Weisfeiler-Leman graph isomorphism.
In this work, we analyze the expressive power of GNNs with random node (RNI)
We prove that these models are universal, a first such result for GNNs not relying on computationally demanding higher-order properties.
arXiv Detail & Related papers (2020-10-02T19:53:05Z) - Distance Encoding: Design Provably More Powerful Neural Networks for
Graph Representation Learning [63.97983530843762]
Graph Neural Networks (GNNs) have achieved great success in graph representation learning.
GNNs generate identical representations for graph substructures that may in fact be very different.
More powerful GNNs, proposed recently by mimicking higher-order tests, are inefficient as they cannot sparsity of underlying graph structure.
We propose Distance Depiction (DE) as a new class of graph representation learning.
arXiv Detail & Related papers (2020-08-31T23:15:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.