Distributed Optimization of Graph Convolutional Network using Subgraph
Variance
- URL: http://arxiv.org/abs/2110.02987v1
- Date: Wed, 6 Oct 2021 18:01:47 GMT
- Title: Distributed Optimization of Graph Convolutional Network using Subgraph
Variance
- Authors: Taige Zhao, Xiangyu Song, Jianxin Li, Wei Luo, Imran Razzak
- Abstract summary: We propose a Graph Augmentation based Distributed GCN framework(GAD)
GAD has two main components, GAD-Partition and GAD-r.
Our framework significantly reduces the communication overhead 50%, improves the convergence speed (2X) and slight gain in accuracy (0.45%) based on minimal redundancy compared to the state-of-the-art methods.
- Score: 8.510726499008204
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, Graph Convolutional Networks (GCNs) have achieved great
success in learning from graph-structured data. With the growing tendency of
graph nodes and edges, GCN training by single processor cannot meet the demand
for time and memory, which led to a boom into distributed GCN training
frameworks research. However, existing distributed GCN training frameworks
require enormous communication costs between processors since multitudes of
dependent nodes and edges information need to be collected and transmitted for
GCN training from other processors. To address this issue, we propose a Graph
Augmentation based Distributed GCN framework(GAD). In particular, GAD has two
main components, GAD-Partition and GAD-Optimizer. We first propose a graph
augmentation-based partition (GAD-Partition) that can divide original graph
into augmented subgraphs to reduce communication by selecting and storing as
few significant nodes of other processors as possible while guaranteeing the
accuracy of the training. In addition, we further design a subgraph
variance-based importance calculation formula and propose a novel weighted
global consensus method, collectively referred to as GAD-Optimizer. This
optimizer adaptively reduces the importance of subgraphs with large variances
for the purpose of reducing the effect of extra variance introduced by
GAD-Partition on distributed GCN training. Extensive experiments on four
large-scale real-world datasets demonstrate that our framework significantly
reduces the communication overhead (50%), improves the convergence speed (2X)
of distributed GCN training, and slight gain in accuracy (0.45%) based on
minimal redundancy compared to the state-of-the-art methods.
Related papers
- MassiveGNN: Efficient Training via Prefetching for Massively Connected Distributed Graphs [11.026326555186333]
This paper develops a parameterized continuous prefetch and eviction scheme on top of the state-of-the-art Amazon DistDGL distributed GNN framework.
It demonstrates about 15-40% improvement in end-to-end training performance on the National Energy Research Scientific Computing Center's (NERSC) Perlmutter supercomputer.
arXiv Detail & Related papers (2024-10-30T05:10:38Z) - Distributed Training of Large Graph Neural Networks with Variable Communication Rates [71.7293735221656]
Training Graph Neural Networks (GNNs) on large graphs presents unique challenges due to the large memory and computing requirements.
Distributed GNN training, where the graph is partitioned across multiple machines, is a common approach to training GNNs on large graphs.
We introduce a variable compression scheme for reducing the communication volume in distributed GNN training without compromising the accuracy of the learned model.
arXiv Detail & Related papers (2024-06-25T14:57:38Z) - T-GAE: Transferable Graph Autoencoder for Network Alignment [79.89704126746204]
T-GAE is a graph autoencoder framework that leverages transferability and stability of GNNs to achieve efficient network alignment without retraining.
Our experiments demonstrate that T-GAE outperforms the state-of-the-art optimization method and the best GNN approach by up to 38.7% and 50.8%, respectively.
arXiv Detail & Related papers (2023-10-05T02:58:29Z) - Communication-Free Distributed GNN Training with Vertex Cut [63.22674903170953]
CoFree-GNN is a novel distributed GNN training framework that significantly speeds up the training process by implementing communication-free training.
We demonstrate that CoFree-GNN speeds up the GNN training process by up to 10 times over the existing state-of-the-art GNN training approaches.
arXiv Detail & Related papers (2023-08-06T21:04:58Z) - ABC: Aggregation before Communication, a Communication Reduction
Framework for Distributed Graph Neural Network Training and Effective
Partition [0.0]
Graph Neural Networks (GNNs) are neural models tailored for graph-structure data and have shown superior performance in learning representations for graph-structured data.
In this paper, we study the communication complexity during distributed GNNs training.
We show that the new partition paradigm is particularly ideal in the case of dynamic graphs where it is infeasible to control the edge placement due to the unknown of the graph-changing process.
arXiv Detail & Related papers (2022-12-11T04:54:01Z) - A Comprehensive Study on Large-Scale Graph Training: Benchmarking and
Rethinking [124.21408098724551]
Large-scale graph training is a notoriously challenging problem for graph neural networks (GNNs)
We present a new ensembling training manner, named EnGCN, to address the existing issues.
Our proposed method has achieved new state-of-the-art (SOTA) performance on large-scale datasets.
arXiv Detail & Related papers (2022-10-14T03:43:05Z) - Comprehensive Graph Gradual Pruning for Sparse Training in Graph Neural
Networks [52.566735716983956]
We propose a graph gradual pruning framework termed CGP to dynamically prune GNNs.
Unlike LTH-based methods, the proposed CGP approach requires no re-training, which significantly reduces the computation costs.
Our proposed strategy greatly improves both training and inference efficiency while matching or even exceeding the accuracy of existing methods.
arXiv Detail & Related papers (2022-07-18T14:23:31Z) - Bi-GCN: Binary Graph Convolutional Network [57.733849700089955]
We propose a Binary Graph Convolutional Network (Bi-GCN), which binarizes both the network parameters and input node features.
Our Bi-GCN can reduce the memory consumption by an average of 30x for both the network parameters and input data, and accelerate the inference speed by an average of 47x.
arXiv Detail & Related papers (2020-10-15T07:26:23Z) - Distributed Training of Graph Convolutional Networks [24.040921719350283]
We show how to make inference in a distributed scenario where the underlying data graph is split among different agents.
We then propose a distributed gradient descent procedure to solve the GCN training problem.
Convergence to stationary solutions of the GCN training problem is also established under mild conditions.
arXiv Detail & Related papers (2020-07-13T10:04:20Z) - An Uncoupled Training Architecture for Large Graph Learning [20.784230322205232]
We present Node2Grids, a flexible uncoupled training framework for embedding graph data into grid-like data.
By ranking each node's influence through degree, Node2Grids selects the most influential first-order as well as second-order neighbors with central node fusion information.
For further improving the efficiency of downstream tasks, a simple CNN-based neural network is employed to capture the significant information from the mapped grid-like data.
arXiv Detail & Related papers (2020-03-21T11:49:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.