IGLU: Efficient GCN Training via Lazy Updates
- URL: http://arxiv.org/abs/2109.13995v1
- Date: Tue, 28 Sep 2021 19:11:00 GMT
- Title: IGLU: Efficient GCN Training via Lazy Updates
- Authors: S Deepak Narayanan, Aditya Sinha, Prateek Jain, Purushottam Kar,
Sundararajan Sellamanickam
- Abstract summary: Graph Convolution Networks (GCN) are used in numerous settings involving a large underlying graph as well as several layers.
Standard SGD-based training scales poorly here since each descent step ends up updating node embeddings for a large portion of the graph.
We introduce a new method IGLU that caches forward-pass embeddings for all nodes at various GCN layers.
- Score: 17.24386142849498
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Graph Convolution Networks (GCN) are used in numerous settings involving a
large underlying graph as well as several layers. Standard SGD-based training
scales poorly here since each descent step ends up updating node embeddings for
a large portion of the graph. Recent methods attempt to remedy this by
sub-sampling the graph which does reduce the compute load, but at the cost of
biased gradients which may offer suboptimal performance. In this work we
introduce a new method IGLU that caches forward-pass embeddings for all nodes
at various GCN layers. This enables IGLU to perform lazy updates that do not
require updating a large number of node embeddings during descent which offers
much faster convergence but does not significantly bias the gradients. Under
standard assumptions such as objective smoothness, IGLU provably converges to a
first-order saddle point. We validate IGLU extensively on a variety of
benchmarks, where it offers up to 1.2% better accuracy despite requiring up to
88% less wall-clock time.
Related papers
- L2G2G: a Scalable Local-to-Global Network Embedding with Graph
Autoencoders [6.945992777272943]
graph representation learning is a popular tool for analysing real-world networks.
GAEs tend to be fairly accurate, but they suffer from scalability issues.
For improved speed, a Local2Global approach was shown to be fast and achieve good accuracy.
Here we propose L2G2G, a Local2Global method which improves GAE accuracy without sacrificing scalability.
arXiv Detail & Related papers (2024-02-02T18:24:37Z) - A Comprehensive Study on Large-Scale Graph Training: Benchmarking and
Rethinking [124.21408098724551]
Large-scale graph training is a notoriously challenging problem for graph neural networks (GNNs)
We present a new ensembling training manner, named EnGCN, to address the existing issues.
Our proposed method has achieved new state-of-the-art (SOTA) performance on large-scale datasets.
arXiv Detail & Related papers (2022-10-14T03:43:05Z) - DOTIN: Dropping Task-Irrelevant Nodes for GNNs [119.17997089267124]
Recent graph learning approaches have introduced the pooling strategy to reduce the size of graphs for learning.
We design a new approach called DOTIN (underlineDrunderlineopping underlineTask-underlineIrrelevant underlineNodes) to reduce the size of graphs.
Our method speeds up GAT by about 50% on graph-level tasks including graph classification and graph edit distance.
arXiv Detail & Related papers (2022-04-28T12:00:39Z) - Gradient Correction beyond Gradient Descent [63.33439072360198]
gradient correction is apparently the most crucial aspect for the training of a neural network.
We introduce a framework (textbfGCGD) to perform gradient correction.
Experiment results show that our gradient correction framework can effectively improve the gradient quality to reduce training epochs by $sim$ 20% and also improve the network performance.
arXiv Detail & Related papers (2022-03-16T01:42:25Z) - Residual Network and Embedding Usage: New Tricks of Node Classification
with Graph Convolutional Networks [0.38233569758620045]
We first summarize some existing effective tricks used in GCNs mini-batch training.
Based on this, two novel tricks named GCN_res Framework and Embedding Usage are proposed.
Experiments on Open Graph Benchmark show that, by combining these techniques, the test accuracy of various GCNs increases by 1.21%2.84%.
arXiv Detail & Related papers (2021-05-18T07:52:51Z) - Combining Label Propagation and Simple Models Out-performs Graph Neural
Networks [52.121819834353865]
We show that for many standard transductive node classification benchmarks, we can exceed or match the performance of state-of-the-art GNNs.
We call this overall procedure Correct and Smooth (C&S)
Our approach exceeds or nearly matches the performance of state-of-the-art GNNs on a wide variety of benchmarks.
arXiv Detail & Related papers (2020-10-27T02:10:52Z) - Sequential Graph Convolutional Network for Active Learning [53.99104862192055]
We propose a novel pool-based Active Learning framework constructed on a sequential Graph Convolution Network (GCN)
With a small number of randomly sampled images as seed labelled examples, we learn the parameters of the graph to distinguish labelled vs unlabelled nodes.
We exploit these characteristics of GCN to select the unlabelled examples which are sufficiently different from labelled ones.
arXiv Detail & Related papers (2020-06-18T00:55:10Z) - L$^2$-GCN: Layer-Wise and Learned Efficient Training of Graph
Convolutional Networks [118.37805042816784]
Graph convolution networks (GCN) are increasingly popular in many applications, yet remain notoriously hard to train over large graph datasets.
We propose a novel efficient layer-wise training framework for GCN (L-GCN), that disentangles feature aggregation and feature transformation during training.
Experiments show that L-GCN is faster than state-of-the-arts by at least an order of magnitude, with a consistent of memory usage not dependent on dataset size.
arXiv Detail & Related papers (2020-03-30T16:37:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.