Tokenized Graph Transformer with Neighborhood Augmentation for Node
Classification in Large Graphs
- URL: http://arxiv.org/abs/2305.12677v1
- Date: Mon, 22 May 2023 03:29:42 GMT
- Title: Tokenized Graph Transformer with Neighborhood Augmentation for Node
Classification in Large Graphs
- Authors: Jinsong Chen, Chang Liu, Kaiyuan Gao, Gaichao Li, Kun He
- Abstract summary: We propose a Neighborhood Aggregation Graph Transformer (NAGphormer) that treats each node as a sequence containing a series of tokens.
Hop2Token aggregates the neighborhood features from different hops into different representations, producing a sequence of token vectors as one input.
In addition, we propose a new data augmentation method called Neighborhood Augmentation (NrAug) based on the output of Hop2Token.
- Score: 11.868008619702277
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph Transformers, emerging as a new architecture for graph representation
learning, suffer from the quadratic complexity on the number of nodes when
handling large graphs. To this end, we propose a Neighborhood Aggregation Graph
Transformer (NAGphormer) that treats each node as a sequence containing a
series of tokens constructed by our proposed Hop2Token module. For each node,
Hop2Token aggregates the neighborhood features from different hops into
different representations, producing a sequence of token vectors as one input.
In this way, NAGphormer could be trained in a mini-batch manner and thus could
scale to large graphs. Moreover, we mathematically show that compared to a
category of advanced Graph Neural Networks (GNNs), called decoupled Graph
Convolutional Networks, NAGphormer could learn more informative node
representations from multi-hop neighborhoods. In addition, we propose a new
data augmentation method called Neighborhood Augmentation (NrAug) based on the
output of Hop2Token that augments simultaneously the features of neighborhoods
from global as well as local views to strengthen the training effect of
NAGphormer. Extensive experiments on benchmark datasets from small to large
demonstrate the superiority of NAGphormer against existing graph Transformers
and mainstream GNNs, and the effectiveness of NrAug for further boosting
NAGphormer.
Related papers
- Leveraging Contrastive Learning for Enhanced Node Representations in Tokenized Graph Transformers [14.123432611346674]
We propose a novel graph Transformer called GCFormer to harness graph information for learning optimal node representations.
GCFormer develops a hybrid token generator to create two types of token sequences, positive and negative, to capture diverse graph information.
A tailored Transformer-based backbone is adopted to learn meaningful node representations from these generated token sequences.
arXiv Detail & Related papers (2024-06-27T15:29:47Z) - NTFormer: A Composite Node Tokenized Graph Transformer for Node Classification [11.451341325579188]
We propose a new graph Transformer called NTFormer to address node classification issues.
New token generator called Node2Par generates various token sequences using different token elements for each node.
Experiments conducted on various benchmark datasets demonstrate the superiority of NTFormer over representative graph Transformers and graph neural networks for node classification.
arXiv Detail & Related papers (2024-06-27T15:16:00Z) - Masked Graph Transformer for Large-Scale Recommendation [56.37903431721977]
We propose an efficient Masked Graph Transformer, named MGFormer, capable of capturing all-pair interactions among nodes with a linear complexity.
Experimental results show the superior performance of our MGFormer, even with a single attention layer.
arXiv Detail & Related papers (2024-05-07T06:00:47Z) - SpikeGraphormer: A High-Performance Graph Transformer with Spiking Graph Attention [1.4126245676224705]
Graph Transformers have emerged as a promising solution to alleviate the inherent limitations of Graph Neural Networks (GNNs)
We propose a novel insight into integrating SNNs with Graph Transformers and design a Spiking Graph Attention (SGA) module.
SpikeGraphormer consistently outperforms existing state-of-the-art approaches across various datasets.
arXiv Detail & Related papers (2024-03-21T03:11:53Z) - Graph Mixture of Experts: Learning on Large-Scale Graphs with Explicit
Diversity Modeling [60.0185734837814]
Graph neural networks (GNNs) have found extensive applications in learning from graph data.
To bolster the generalization capacity of GNNs, it has become customary to augment training graph structures with techniques like graph augmentations.
This study introduces the concept of Mixture-of-Experts (MoE) to GNNs, with the aim of augmenting their capacity to adapt to a diverse range of training graph structures.
arXiv Detail & Related papers (2023-04-06T01:09:36Z) - Pure Transformers are Powerful Graph Learners [51.36884247453605]
We show that standard Transformers without graph-specific modifications can lead to promising results in graph learning both in theory and practice.
We prove that this approach is theoretically at least as expressive as an invariant graph network (2-IGN) composed of equivariant linear layers.
Our method coined Tokenized Graph Transformer (TokenGT) achieves significantly better results compared to GNN baselines and competitive results.
arXiv Detail & Related papers (2022-07-06T08:13:06Z) - NAGphormer: Neighborhood Aggregation Graph Transformer for Node
Classification in Large Graphs [10.149586598073421]
We propose a Neighborhood Aggregation Graph Transformer (NAGphormer) that is scalable to large graphs with millions of nodes.
NAGphormer constructs tokens for each node by a neighborhood aggregation module called Hop2Token.
We conduct extensive experiments on various popular benchmarks, including six small datasets and three large datasets.
arXiv Detail & Related papers (2022-06-10T07:23:51Z) - Neighbor2Seq: Deep Learning on Massive Graphs by Transforming Neighbors
to Sequences [55.329402218608365]
We propose the Neighbor2Seq to transform the hierarchical neighborhood of each node into a sequence.
We evaluate our method on a massive graph with more than 111 million nodes and 1.6 billion edges.
Results show that our proposed method is scalable to massive graphs and achieves superior performance across massive and medium-scale graphs.
arXiv Detail & Related papers (2022-02-07T16:38:36Z) - Graph Neural Networks with Learnable Structural and Positional
Representations [83.24058411666483]
A major issue with arbitrary graphs is the absence of canonical positional information of nodes.
We introduce Positional nodes (PE) of nodes, and inject it into the input layer, like in Transformers.
We observe a performance increase for molecular datasets, from 2.87% up to 64.14% when considering learnable PE for both GNN classes.
arXiv Detail & Related papers (2021-10-15T05:59:15Z) - Bilinear Graph Neural Network with Neighbor Interactions [106.80781016591577]
Graph Neural Network (GNN) is a powerful model to learn representations and make predictions on graph data.
We propose a new graph convolution operator, which augments the weighted sum with pairwise interactions of the representations of neighbor nodes.
We term this framework as Bilinear Graph Neural Network (BGNN), which improves GNN representation ability with bilinear interactions between neighbor nodes.
arXiv Detail & Related papers (2020-02-10T06:43:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.