Building Shortcuts between Distant Nodes with Biaffine Mapping for Graph
Convolutional Networks
- URL: http://arxiv.org/abs/2302.08727v2
- Date: Mon, 20 Feb 2023 02:40:13 GMT
- Title: Building Shortcuts between Distant Nodes with Biaffine Mapping for Graph
Convolutional Networks
- Authors: Acong Zhang and Jincheng Huang and Ping Li and Kai Zhang
- Abstract summary: We introduce Biaffine technique to improve the expressiveness of graph convolutional networks with a shallow architecture.
Our method is to learn direct dependency on long-distance neighbors for nodes, with which only one-hop message passing is capable of capturing rich information for node representation.
- Score: 18.160610500658183
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multiple recent studies show a paradox in graph convolutional networks
(GCNs), that is, shallow architectures limit the capability of learning
information from high-order neighbors, while deep architectures suffer from
over-smoothing or over-squashing. To enjoy the simplicity of shallow
architectures and overcome their limits of neighborhood extension, in this
work, we introduce Biaffine technique to improve the expressiveness of graph
convolutional networks with a shallow architecture. The core design of our
method is to learn direct dependency on long-distance neighbors for nodes, with
which only one-hop message passing is capable of capturing rich information for
node representation. Besides, we propose a multi-view contrastive learning
method to exploit the representations learned from long-distance dependencies.
Extensive experiments on nine graph benchmark datasets suggest that the shallow
biaffine graph convolutional networks (BAGCN) significantly outperforms
state-of-the-art GCNs (with deep or shallow architectures) on semi-supervised
node classification. We further verify the effectiveness of biaffine design in
node representation learning and the performance consistency on different sizes
of training data.
Related papers
- Contrastive Learning for Non-Local Graphs with Multi-Resolution
Structural Views [1.4445779250002606]
We propose a novel multiview contrastive learning approach that integrates diffusion filters on graphs.
By incorporating multiple graph views as augmentations, our method captures the structural equivalence in heterophilic graphs.
arXiv Detail & Related papers (2023-08-19T17:42:02Z) - NodeFormer: A Scalable Graph Structure Learning Transformer for Node
Classification [70.51126383984555]
We introduce a novel all-pair message passing scheme for efficiently propagating node signals between arbitrary nodes.
The efficient computation is enabled by a kernerlized Gumbel-Softmax operator.
Experiments demonstrate the promising efficacy of the method in various tasks including node classification on graphs.
arXiv Detail & Related papers (2023-06-14T09:21:15Z) - Uniting Heterogeneity, Inductiveness, and Efficiency for Graph
Representation Learning [68.97378785686723]
graph neural networks (GNNs) have greatly advanced the performance of node representation learning on graphs.
A majority class of GNNs are only designed for homogeneous graphs, leading to inferior adaptivity to the more informative heterogeneous graphs.
We propose a novel inductive, meta path-free message passing scheme that packs up heterogeneous node features with their associated edges from both low- and high-order neighbor nodes.
arXiv Detail & Related papers (2021-04-04T23:31:39Z) - RAN-GNNs: breaking the capacity limits of graph neural networks [43.66682619000099]
Graph neural networks have become a staple in problems addressing learning and analysis of data defined over graphs.
Recent works attribute this to the need to consider multiple neighborhood sizes at the same time and adaptively tune them.
We show that employing a randomly-wired architecture can be a more effective way to increase the capacity of the network and obtain richer representations.
arXiv Detail & Related papers (2021-03-29T12:34:36Z) - Multi-Level Attention Pooling for Graph Neural Networks: Unifying Graph
Representations with Multiple Localities [4.142375560633827]
Graph neural networks (GNNs) have been widely used to learn vector representation of graph-structured data.
A potential cause is that deep GNN models tend to lose the nodes' local information through many message passing steps.
We propose a multi-level attention pooling architecture to solve this so-called oversmoothing problem.
arXiv Detail & Related papers (2021-03-02T05:58:12Z) - Spatio-Temporal Inception Graph Convolutional Networks for
Skeleton-Based Action Recognition [126.51241919472356]
We design a simple and highly modularized graph convolutional network architecture for skeleton-based action recognition.
Our network is constructed by repeating a building block that aggregates multi-granularity information from both the spatial and temporal paths.
arXiv Detail & Related papers (2020-11-26T14:43:04Z) - Graph Fairing Convolutional Networks for Anomaly Detection [7.070726553564701]
We introduce a graph convolutional network with skip connections for semi-supervised anomaly detection.
The effectiveness of our model is demonstrated through extensive experiments on five benchmark datasets.
arXiv Detail & Related papers (2020-10-20T13:45:47Z) - Representation Learning of Graphs Using Graph Convolutional Multilayer
Networks Based on Motifs [17.823543937167848]
mGCMN is a novel framework which utilizes node feature information and the higher order local structure of the graph.
It will greatly improve the learning efficiency of the graph neural network and promote a brand-new learning mode establishment.
arXiv Detail & Related papers (2020-07-31T04:18:20Z) - Towards Deeper Graph Neural Networks [63.46470695525957]
Graph convolutions perform neighborhood aggregation and represent one of the most important graph operations.
Several recent studies attribute this performance deterioration to the over-smoothing issue.
We propose Deep Adaptive Graph Neural Network (DAGNN) to adaptively incorporate information from large receptive fields.
arXiv Detail & Related papers (2020-07-18T01:11:14Z) - Geometrically Principled Connections in Graph Neural Networks [66.51286736506658]
We argue geometry should remain the primary driving force behind innovation in the emerging field of geometric deep learning.
We relate graph neural networks to widely successful computer graphics and data approximation models: radial basis functions (RBFs)
We introduce affine skip connections, a novel building block formed by combining a fully connected layer with any graph convolution operator.
arXiv Detail & Related papers (2020-04-06T13:25:46Z) - Tensor Graph Convolutional Networks for Multi-relational and Robust
Learning [74.05478502080658]
This paper introduces a tensor-graph convolutional network (TGCN) for scalable semi-supervised learning (SSL) from data associated with a collection of graphs, that are represented by a tensor.
The proposed architecture achieves markedly improved performance relative to standard GCNs, copes with state-of-the-art adversarial attacks, and leads to remarkable SSL performance over protein-to-protein interaction networks.
arXiv Detail & Related papers (2020-03-15T02:33:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.