Graph Attention Retrospective
- URL: http://arxiv.org/abs/2202.13060v5
- Date: Mon, 22 May 2023 01:48:07 GMT
- Title: Graph Attention Retrospective
- Authors: Kimon Fountoulakis, Amit Levi, Shenghao Yang, Aseem Baranwal, Aukosh
Jagannath
- Abstract summary: Graph-based learning is a rapidly growing sub-field of machine learning with applications in social networks, citation networks, and bioinformatics.
In this paper, we theoretically study the behaviour of graph attention networks.
We show that in an "easy" regime, where the distance between the means of the Gaussians is large enough, graph attention is able to distinguish inter-class from intra-class edges.
In the "hard" regime, we show that every attention mechanism fails to distinguish intra-class from inter-class edges.
- Score: 14.52271219759284
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph-based learning is a rapidly growing sub-field of machine learning with
applications in social networks, citation networks, and bioinformatics. One of
the most popular models is graph attention networks. They were introduced to
allow a node to aggregate information from features of neighbor nodes in a
non-uniform way, in contrast to simple graph convolution which does not
distinguish the neighbors of a node. In this paper, we theoretically study the
behaviour of graph attention networks. We prove multiple results on the
performance of the graph attention mechanism for the problem of node
classification for a contextual stochastic block model. Here, the node features
are obtained from a mixture of Gaussians and the edges from a stochastic block
model. We show that in an "easy" regime, where the distance between the means
of the Gaussians is large enough, graph attention is able to distinguish
inter-class from intra-class edges. Thus it maintains the weights of important
edges and significantly reduces the weights of unimportant edges. Consequently,
we show that this implies perfect node classification. In the "hard" regime, we
show that every attention mechanism fails to distinguish intra-class from
inter-class edges. In addition, we show that graph attention convolution cannot
(almost) perfectly classify the nodes even if intra-class edges could be
separated from inter-class edges. Beyond perfect node classification, we
provide a positive result on graph attention's robustness against structural
noise in the graph. In particular, our robustness result implies that graph
attention can be strictly better than both the simple graph convolution and the
best linear classifier of node features. We evaluate our theoretical results on
synthetic and real-world data.
Related papers
- Saliency-Aware Regularized Graph Neural Network [39.82009838086267]
We propose the Saliency-Aware Regularized Graph Neural Network (SAR-GNN) for graph classification.
We first estimate the global node saliency by measuring the semantic similarity between the compact graph representation and node features.
Then the learned saliency distribution is leveraged to regularize the neighborhood aggregation of the backbone.
arXiv Detail & Related papers (2024-01-01T13:44:16Z) - NodeFormer: A Scalable Graph Structure Learning Transformer for Node
Classification [70.51126383984555]
We introduce a novel all-pair message passing scheme for efficiently propagating node signals between arbitrary nodes.
The efficient computation is enabled by a kernerlized Gumbel-Softmax operator.
Experiments demonstrate the promising efficacy of the method in various tasks including node classification on graphs.
arXiv Detail & Related papers (2023-06-14T09:21:15Z) - On Classification Thresholds for Graph Attention with Edge Features [26.01769042481568]
We analyze, theoretically and empirically, graph attention networks and their ability of correctly labelling nodes in a classic classification task.
We consider a general graph attention mechanism that takes random edge features as input to determine the attention coefficients.
arXiv Detail & Related papers (2022-10-18T17:32:18Z) - Neighborhood Random Walk Graph Sampling for Regularized Bayesian Graph
Convolutional Neural Networks [0.6236890292833384]
In this paper, we propose a novel algorithm called Bayesian Graph Convolutional Network using Neighborhood Random Walk Sampling (BGCN-NRWS)
BGCN-NRWS uses a Markov Chain Monte Carlo (MCMC) based graph sampling algorithm utilizing graph structure, reduces overfitting by using a variational inference layer, and yields consistently competitive classification results compared to the state-of-the-art in semi-supervised node classification.
arXiv Detail & Related papers (2021-12-14T20:58:27Z) - Graph Kernel Neural Networks [53.91024360329517]
We propose to use graph kernels, i.e. kernel functions that compute an inner product on graphs, to extend the standard convolution operator to the graph domain.
This allows us to define an entirely structural model that does not require computing the embedding of the input graph.
Our architecture allows to plug-in any type of graph kernels and has the added benefit of providing some interpretability.
arXiv Detail & Related papers (2021-12-14T14:48:08Z) - Graph Neural Networks with Feature and Structure Aware Random Walk [7.143879014059894]
We show that in typical heterphilous graphs, the edges may be directed, and whether to treat the edges as is or simply make them undirected greatly affects the performance of the GNN models.
We develop a model that adaptively learns the directionality of the graph, and exploits the underlying long-distance correlations between nodes.
arXiv Detail & Related papers (2021-11-19T08:54:21Z) - Edge but not Least: Cross-View Graph Pooling [76.71497833616024]
This paper presents a cross-view graph pooling (Co-Pooling) method to better exploit crucial graph structure information.
Through cross-view interaction, edge-view pooling and node-view pooling seamlessly reinforce each other to learn more informative graph-level representations.
arXiv Detail & Related papers (2021-09-24T08:01:23Z) - Graph Decoupling Attention Markov Networks for Semi-supervised Graph
Node Classification [38.52231889960877]
Graph neural networks (GNN) have been ubiquitous in graph learning tasks such as node classification.
In this paper, we consider the label dependency of graph nodes and propose a decoupling attention mechanism to learn both hard and soft attention.
arXiv Detail & Related papers (2021-04-28T11:44:13Z) - Higher-Order Attribute-Enhancing Heterogeneous Graph Neural Networks [67.25782890241496]
We propose a higher-order Attribute-Enhancing Graph Neural Network (HAEGNN) for heterogeneous network representation learning.
HAEGNN simultaneously incorporates meta-paths and meta-graphs for rich, heterogeneous semantics.
It shows superior performance against the state-of-the-art methods in node classification, node clustering, and visualization.
arXiv Detail & Related papers (2021-04-16T04:56:38Z) - Line Graph Neural Networks for Link Prediction [71.00689542259052]
We consider the graph link prediction task, which is a classic graph analytical problem with many real-world applications.
In this formalism, a link prediction problem is converted to a graph classification task.
We propose to seek a radically different and novel path by making use of the line graphs in graph theory.
In particular, each node in a line graph corresponds to a unique edge in the original graph. Therefore, link prediction problems in the original graph can be equivalently solved as a node classification problem in its corresponding line graph, instead of a graph classification task.
arXiv Detail & Related papers (2020-10-20T05:54:31Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.