Adaptive Multi-Neighborhood Attention based Transformer for Graph
Representation Learning
- URL: http://arxiv.org/abs/2211.07970v1
- Date: Tue, 15 Nov 2022 08:12:44 GMT
- Title: Adaptive Multi-Neighborhood Attention based Transformer for Graph
Representation Learning
- Authors: Gaichao Li, Jinsong Chen, Kun He
- Abstract summary: We propose an adaptive graph Transformer termed Multi-Neighborhood Attention based Graph Transformer (MNA-GT)
MNA-GT captures the graph structural information for each node from the multi-neighborhood attention mechanism adaptively.
Experiments are conducted on a variety of graph benchmarks, and the empirical results show that MNA-GT outperforms many strong baselines.
- Score: 11.407118196728943
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: By incorporating the graph structural information into Transformers, graph
Transformers have exhibited promising performance for graph representation
learning in recent years. Existing graph Transformers leverage specific
strategies, such as Laplacian eigenvectors and shortest paths of the node
pairs, to preserve the structural features of nodes and feed them into the
vanilla Transformer to learn the representations of nodes. It is hard for such
predefined rules to extract informative graph structural features for arbitrary
graphs whose topology structure varies greatly, limiting the learning capacity
of the models. To this end, we propose an adaptive graph Transformer, termed
Multi-Neighborhood Attention based Graph Transformer (MNA-GT), which captures
the graph structural information for each node from the multi-neighborhood
attention mechanism adaptively. By defining the input to perform scaled-dot
product as an attention kernel, MNA-GT constructs multiple attention kernels
based on different hops of neighborhoods such that each attention kernel can
capture specific graph structural information of the corresponding neighborhood
for each node pair. In this way, MNA-GT can preserve the graph structural
information efficiently by incorporating node representations learned by
different attention kernels. MNA-GT further employs an attention layer to learn
the importance of different attention kernels to enable the model to adaptively
capture the graph structural information for different nodes. Extensive
experiments are conducted on a variety of graph benchmarks, and the empirical
results show that MNA-GT outperforms many strong baselines.
Related papers
- InstructG2I: Synthesizing Images from Multimodal Attributed Graphs [50.852150521561676]
We propose a graph context-conditioned diffusion model called InstructG2I.
InstructG2I first exploits the graph structure and multimodal information to conduct informative neighbor sampling.
A Graph-QFormer encoder adaptively encodes the graph nodes into an auxiliary set of graph prompts to guide the denoising process.
arXiv Detail & Related papers (2024-10-09T17:56:15Z) - A Pure Transformer Pretraining Framework on Text-attributed Graphs [50.833130854272774]
We introduce a feature-centric pretraining perspective by treating graph structure as a prior.
Our framework, Graph Sequence Pretraining with Transformer (GSPT), samples node contexts through random walks.
GSPT can be easily adapted to both node classification and link prediction, demonstrating promising empirical success on various datasets.
arXiv Detail & Related papers (2024-06-19T22:30:08Z) - Technical Report: The Graph Spectral Token -- Enhancing Graph Transformers with Spectral Information [0.8184895397419141]
Graph Transformers have emerged as a powerful alternative to Message-Passing Graph Neural Networks (MP-GNNs)
We propose the Graph Spectral Token, a novel approach to directly encode graph spectral information.
We benchmark the effectiveness of our approach by enhancing two existing graph transformers, GraphTrans and SubFormer.
arXiv Detail & Related papers (2024-04-08T15:24:20Z) - SignGT: Signed Attention-based Graph Transformer for Graph
Representation Learning [15.248591535696146]
We propose a Signed Attention-based Graph Transformer (SignGT) to adaptively capture various frequency information from the graphs.
Specifically, SignGT develops a new signed self-attention mechanism (SignSA) that produces signed attention values according to the semantic relevance of node pairs.
arXiv Detail & Related papers (2023-10-17T06:42:11Z) - SGFormer: Simplifying and Empowering Transformers for Large-Graph Representations [75.71298846760303]
We show that a one-layer attention can bring up surprisingly competitive performance across node property prediction benchmarks.
We frame the proposed scheme as Simplified Graph Transformers (SGFormer), which is empowered by a simple attention model.
We believe the proposed methodology alone enlightens a new technical path of independent interest for building Transformers on large graphs.
arXiv Detail & Related papers (2023-06-19T08:03:25Z) - Graph Propagation Transformer for Graph Representation Learning [36.01189696668657]
We propose a new attention mechanism called Graph Propagation Attention (GPA)
It explicitly passes the information among nodes and edges in three ways, i.e. node-to-node, node-to-edge, and edge-to-node.
We show that our method outperforms many state-of-the-art transformer-based graph models with better performance.
arXiv Detail & Related papers (2023-05-19T04:42:58Z) - Diffusing Graph Attention [15.013509382069046]
We develop a new model for Graph Transformers that integrates the arbitrary graph structure into the architecture.
GD learns to extract structural and positional relationships between distant nodes in the graph, which it then uses to direct the Transformer's attention and node representation.
Experiments on eight benchmarks show Graph diffuser to be a highly competitive model, outperforming the state-of-the-art in a diverse set of domains.
arXiv Detail & Related papers (2023-03-01T16:11:05Z) - SHGNN: Structure-Aware Heterogeneous Graph Neural Network [77.78459918119536]
This paper proposes a novel Structure-Aware Heterogeneous Graph Neural Network (SHGNN) to address the above limitations.
We first utilize a feature propagation module to capture the local structure information of intermediate nodes in the meta-path.
Next, we use a tree-attention aggregator to incorporate the graph structure information into the aggregation module on the meta-path.
Finally, we leverage a meta-path aggregator to fuse the information aggregated from different meta-paths.
arXiv Detail & Related papers (2021-12-12T14:18:18Z) - Graph Neural Networks with Learnable Structural and Positional
Representations [83.24058411666483]
A major issue with arbitrary graphs is the absence of canonical positional information of nodes.
We introduce Positional nodes (PE) of nodes, and inject it into the input layer, like in Transformers.
We observe a performance increase for molecular datasets, from 2.87% up to 64.14% when considering learnable PE for both GNN classes.
arXiv Detail & Related papers (2021-10-15T05:59:15Z) - Spectral Graph Convolutional Networks With Lifting-based Adaptive Graph
Wavelets [81.63035727821145]
Spectral graph convolutional networks (SGCNs) have been attracting increasing attention in graph representation learning.
We propose a novel class of spectral graph convolutional networks that implement graph convolutions with adaptive graph wavelets.
arXiv Detail & Related papers (2021-08-03T17:57:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.