SEA: Graph Shell Attention in Graph Neural Networks
- URL: http://arxiv.org/abs/2110.10674v1
- Date: Wed, 20 Oct 2021 17:32:08 GMT
- Title: SEA: Graph Shell Attention in Graph Neural Networks
- Authors: Christian M.M. Frey, Yunpu Ma, Matthias Schubert
- Abstract summary: A common issue in Graph Neural Networks (GNNs) is known as over-smoothing.
In our work, we relax the GNN architecture by means of implementing a routing. Specifically, the nodes' representations are routed to dedicated experts.
We call this procedure Graph Shell Attention (SEA), where experts process different subgraphs in a transformer-motivated fashion.
- Score: 8.565134944225491
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: A common issue in Graph Neural Networks (GNNs) is known as over-smoothing. By
increasing the number of iterations within the message-passing of GNNs, the
nodes' representations of the input graph align with each other and become
indiscernible. Recently, it has been shown that increasing a model's complexity
by integrating an attention mechanism yields more expressive architectures.
This is majorly contributed to steering the nodes' representations only towards
nodes that are more informative than others. Transformer models in combination
with GNNs result in architectures including Graph Transformer Layers (GTL),
where layers are entirely based on the attention operation. However, the
calculation of a node's representation is still restricted to the computational
working flow of a GNN. In our work, we relax the GNN architecture by means of
implementing a routing heuristic. Specifically, the nodes' representations are
routed to dedicated experts. Each expert calculates the representations
according to their respective GNN workflow. The definitions of distinguishable
GNNs result from k-localized views starting from the central node. We call this
procedure Graph Shell Attention (SEA), where experts process different
subgraphs in a transformer-motivated fashion. Intuitively, by increasing the
number of experts, the models gain in expressiveness such that a node's
representation is solely based on nodes that are located within the receptive
field of an expert. We evaluate our architecture on various benchmark datasets
showing competitive results compared to state-of-the-art models.
Related papers
- SF-GNN: Self Filter for Message Lossless Propagation in Deep Graph Neural Network [38.669815079957566]
Graph Neural Network (GNN) with the main idea of encoding graph structure information of graphs by propagation and aggregation has developed rapidly.
It achieved excellent performance in representation learning of multiple types of graphs such as homogeneous graphs, heterogeneous graphs, and more complex graphs like knowledge graphs.
For the phenomenon of performance degradation in deep GNNs, we propose a new perspective.
arXiv Detail & Related papers (2024-07-03T02:40:39Z) - Degree-based stratification of nodes in Graph Neural Networks [66.17149106033126]
We modify the Graph Neural Network (GNN) architecture so that the weight matrices are learned, separately, for the nodes in each group.
This simple-to-implement modification seems to improve performance across datasets and GNN methods.
arXiv Detail & Related papers (2023-12-16T14:09:23Z) - Seq-HGNN: Learning Sequential Node Representation on Heterogeneous Graph [57.2953563124339]
We propose a novel heterogeneous graph neural network with sequential node representation, namely Seq-HGNN.
We conduct extensive experiments on four widely used datasets from Heterogeneous Graph Benchmark (HGB) and Open Graph Benchmark (OGB)
arXiv Detail & Related papers (2023-05-18T07:27:18Z) - Graph Neural Networks with Feature and Structure Aware Random Walk [7.143879014059894]
We show that in typical heterphilous graphs, the edges may be directed, and whether to treat the edges as is or simply make them undirected greatly affects the performance of the GNN models.
We develop a model that adaptively learns the directionality of the graph, and exploits the underlying long-distance correlations between nodes.
arXiv Detail & Related papers (2021-11-19T08:54:21Z) - Graph Neural Networks with Learnable Structural and Positional
Representations [83.24058411666483]
A major issue with arbitrary graphs is the absence of canonical positional information of nodes.
We introduce Positional nodes (PE) of nodes, and inject it into the input layer, like in Transformers.
We observe a performance increase for molecular datasets, from 2.87% up to 64.14% when considering learnable PE for both GNN classes.
arXiv Detail & Related papers (2021-10-15T05:59:15Z) - Feature Correlation Aggregation: on the Path to Better Graph Neural
Networks [37.79964911718766]
Prior to the introduction of Graph Neural Networks (GNNs), modeling and analyzing irregular data, particularly graphs, was thought to be the Achilles' heel of deep learning.
This paper introduces a central node permutation variant function through a frustratingly simple and innocent-looking modification to the core operation of a GNN.
A tangible boost in performance of the model is observed where the model surpasses previous state-of-the-art results by a significant margin while employing fewer parameters.
arXiv Detail & Related papers (2021-09-20T05:04:26Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - Multi-grained Semantics-aware Graph Neural Networks [13.720544777078642]
Graph Neural Networks (GNNs) are powerful techniques in representation learning for graphs.
This work proposes a unified model, AdamGNN, to interactively learn node and graph representations.
Experiments on 14 real-world graph datasets show that AdamGNN can significantly outperform 17 competing models on both node- and graph-wise tasks.
arXiv Detail & Related papers (2020-10-01T07:52:06Z) - Distance Encoding: Design Provably More Powerful Neural Networks for
Graph Representation Learning [63.97983530843762]
Graph Neural Networks (GNNs) have achieved great success in graph representation learning.
GNNs generate identical representations for graph substructures that may in fact be very different.
More powerful GNNs, proposed recently by mimicking higher-order tests, are inefficient as they cannot sparsity of underlying graph structure.
We propose Distance Depiction (DE) as a new class of graph representation learning.
arXiv Detail & Related papers (2020-08-31T23:15:40Z) - Graph Neural Networks: Architectures, Stability and Transferability [176.3960927323358]
Graph Neural Networks (GNNs) are information processing architectures for signals supported on graphs.
They are generalizations of convolutional neural networks (CNNs) in which individual layers contain banks of graph convolutional filters.
arXiv Detail & Related papers (2020-08-04T18:57:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.