SUREL+: Moving from Walks to Sets for Scalable Subgraph-based Graph
Representation Learning
- URL: http://arxiv.org/abs/2303.03379v3
- Date: Wed, 27 Dec 2023 04:59:38 GMT
- Title: SUREL+: Moving from Walks to Sets for Scalable Subgraph-based Graph
Representation Learning
- Authors: Haoteng Yin, Muhan Zhang, Jianguo Wang, Pan Li
- Abstract summary: Subgraph-based graph representation learning (SGRL) has emerged as a powerful tool in many prediction tasks on graphs.
We propose a novel framework SUREL+ that upgrades SUREL by using node sets instead of walks to represent subgraphs.
- Score: 30.77279782495266
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Subgraph-based graph representation learning (SGRL) has recently emerged as a
powerful tool in many prediction tasks on graphs due to its advantages in model
expressiveness and generalization ability. Most previous SGRL models face
computational challenges associated with the high cost of subgraph extraction
for each training or test query. Recently, SUREL was proposed to accelerate
SGRL, which samples random walks offline and joins these walks online as a
proxy of subgraph for representation learning. Thanks to the reusability of
sampled walks across different queries, SUREL achieves state-of-the-art
performance in terms of scalability and prediction accuracy. However, SUREL
still suffers from high computational overhead caused by node duplication in
sampled walks. In this work, we propose a novel framework SUREL+ that upgrades
SUREL by using node sets instead of walks to represent subgraphs. This
set-based representation eliminates repeated nodes by definition but can also
be irregular in size. To address this issue, we design a customized sparse data
structure to efficiently store and access node sets and provide a specialized
operator to join them in parallel batches. SUREL+ is modularized to support
multiple types of set samplers, structural features, and neural encoders to
complement the structural information loss after the reduction from walks to
sets. Extensive experiments have been performed to validate SUREL+ in the
prediction tasks of links, relation types, and higher-order patterns. SUREL+
achieves 3-11$\times$ speedups of SUREL while maintaining comparable or even
better prediction performance; compared to other SGRL baselines, SUREL+
achieves $\sim$20$\times$ speedups and significantly improves the prediction
accuracy.
Related papers
- Efficient Heterogeneous Graph Learning via Random Projection [58.4138636866903]
Heterogeneous Graph Neural Networks (HGNNs) are powerful tools for deep learning on heterogeneous graphs.
Recent pre-computation-based HGNNs use one-time message passing to transform a heterogeneous graph into regular-shaped tensors.
We propose a hybrid pre-computation-based HGNN, named Random Projection Heterogeneous Graph Neural Network (RpHGNN)
arXiv Detail & Related papers (2023-10-23T01:25:44Z) - Sparsity exploitation via discovering graphical models in multi-variate
time-series forecasting [1.2762298148425795]
We propose a decoupled training method, which includes a graph generating module and a GNNs forecasting module.
First, we use Graphical Lasso (or GraphLASSO) to directly exploit the sparsity pattern from data to build graph structures.
Second, we fit these graph structures and the input data into a Graph Convolutional Recurrent Network (GCRN) to train a forecasting model.
arXiv Detail & Related papers (2023-06-29T16:48:00Z) - A Comprehensive Study on Large-Scale Graph Training: Benchmarking and
Rethinking [124.21408098724551]
Large-scale graph training is a notoriously challenging problem for graph neural networks (GNNs)
We present a new ensembling training manner, named EnGCN, to address the existing issues.
Our proposed method has achieved new state-of-the-art (SOTA) performance on large-scale datasets.
arXiv Detail & Related papers (2022-10-14T03:43:05Z) - From Spectral Graph Convolutions to Large Scale Graph Convolutional
Networks [0.0]
Graph Convolutional Networks (GCNs) have been shown to be a powerful concept that has been successfully applied to a large variety of tasks.
We study the theory that paved the way to the definition of GCN, including related parts of classical graph theory.
arXiv Detail & Related papers (2022-07-12T16:57:08Z) - A Robust Stacking Framework for Training Deep Graph Models with
Multifaceted Node Features [61.92791503017341]
Graph Neural Networks (GNNs) with numerical node features and graph structure as inputs have demonstrated superior performance on various supervised learning tasks with graph data.
The best models for such data types in most standard supervised learning settings with IID (non-graph) data are not easily incorporated into a GNN.
Here we propose a robust stacking framework that fuses graph-aware propagation with arbitrary models intended for IID data.
arXiv Detail & Related papers (2022-06-16T22:46:33Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - Algorithm and System Co-design for Efficient Subgraph-based Graph
Representation Learning [16.170895692951]
Subgraph-based graph representation learning (SGRL) has been recently proposed to deal with some fundamental challenges encountered by canonical graph neural networks (GNNs)
We propose a novel framework SUREL for scalable SGRL by co-designing the learning algorithm and its system support.
arXiv Detail & Related papers (2022-02-28T04:29:22Z) - Scalable Graph Neural Networks for Heterogeneous Graphs [12.44278942365518]
Graph neural networks (GNNs) are a popular class of parametric model for learning over graph-structured data.
Recent work has argued that GNNs primarily use the graph for feature smoothing, and have shown competitive results on benchmark tasks.
In this work, we ask whether these results can be extended to heterogeneous graphs, which encode multiple types of relationship between different entities.
arXiv Detail & Related papers (2020-11-19T06:03:35Z) - Robust Optimization as Data Augmentation for Large-scale Graphs [117.2376815614148]
We propose FLAG (Free Large-scale Adversarial Augmentation on Graphs), which iteratively augments node features with gradient-based adversarial perturbations during training.
FLAG is a general-purpose approach for graph data, which universally works in node classification, link prediction, and graph classification tasks.
arXiv Detail & Related papers (2020-10-19T21:51:47Z) - Stochastic Graph Recurrent Neural Network [6.656993023468793]
We propose SGRNN, a novel neural architecture that applies latent variables to simultaneously capture evolution in node attributes and topology.
Specifically, deterministic states are separated from states in the iterative process to suppress mutual interference.
Experiments on real-world datasets demonstrate the effectiveness of the proposed model.
arXiv Detail & Related papers (2020-09-01T16:14:30Z) - Graph Ordering: Towards the Optimal by Learning [69.72656588714155]
Graph representation learning has achieved a remarkable success in many graph-based applications, such as node classification, prediction, and community detection.
However, for some kind of graph applications, such as graph compression and edge partition, it is very hard to reduce them to some graph representation learning tasks.
In this paper, we propose to attack the graph ordering problem behind such applications by a novel learning approach.
arXiv Detail & Related papers (2020-01-18T09:14:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.