TC-GNN: Bridging Sparse GNN Computation and Dense Tensor Cores on GPUs
- URL: http://arxiv.org/abs/2112.02052v4
- Date: Wed, 31 May 2023 19:24:58 GMT
- Title: TC-GNN: Bridging Sparse GNN Computation and Dense Tensor Cores on GPUs
- Authors: Yuke Wang, Boyuan Feng, Zheng Wang, Guyue Huang, Yufei Ding
- Abstract summary: We propose TC-GNN, the first GNN framework based on GPU Core Units (TCUs)
The core idea is to reconcile the "Sparse" GNN with the high-performance "Dense" TCUs.
Rigorous experiments show an average of 1.70 speedup over the state-of-the-art DGL framework.
- Score: 21.63854538768414
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, graph neural networks (GNNs), as the backbone of graph-based
machine learning, demonstrate great success in various domains (e.g.,
e-commerce). However, the performance of GNNs is usually unsatisfactory due to
the highly sparse and irregular graph-based operations. To this end, we propose
TC-GNN, the first GNN acceleration framework based on GPU Tensor Core Units
(TCUs). The core idea is to reconcile the "Sparse" GNN computation with the
high-performance "Dense" TCUs. Specifically, we conduct an in-depth analysis of
the sparse operations in mainstream GNN computing frameworks. We introduce a
novel sparse graph translation technique to facilitate TCU processing of the
sparse GNN workload. We implement an effective CUDA core and TCU collaboration
design to fully utilize GPU resources. We integrate TC-GNN with the PyTorch
framework for high programmability. Rigorous experiments show an average of
1.70X speedup over the state-of-the-art DGL framework across various models and
datasets.
Related papers
- GCV-Turbo: End-to-end Acceleration of GNN-based Computer Vision Tasks on FPGA [3.2507129535290926]
Graph neural networks (GNNs) have recently empowered various novel computer vision (CV) tasks.
This paper introduces GCV-Turbo, a domain-specific accelerator on FPGA for end-to-end acceleration of GNN-based CV tasks.
arXiv Detail & Related papers (2024-04-10T17:41:41Z) - GeoT: Tensor Centric Library for Graph Neural Network via Efficient Segment Reduction on GPU [8.15747734801831]
We introduce GeoT, a cutting-edge tensor-centric library designed specifically for Graph Neural Networks (GNNs)
GeoT debuts innovative parallel algorithms that not only introduce new design principles but also expand the available design space.
GeoT marks a considerable advancement by showcasing an average operator speedup of 1.80x and an end-to-end speedup of 1.68x.
arXiv Detail & Related papers (2024-04-03T19:03:15Z) - BitGNN: Unleashing the Performance Potential of Binary Graph Neural
Networks on GPUs [19.254040098787893]
Recent studies have shown that Binary Graph Neural Networks (GNNs) are promising for saving computations of GNNs through binarized tensors.
This work redesigns the binary GNN inference from the efficiency perspective.
Results on real-world graphs with GCNs, GraphSAGE, and GraphSAINT show that the proposed techniques outperform state-of-the-art binary GNN implementations by 8-22X with the same accuracy maintained.
arXiv Detail & Related papers (2023-05-04T03:20:25Z) - Distributed Graph Neural Network Training: A Survey [51.77035975191926]
Graph neural networks (GNNs) are a type of deep learning models that are trained on graphs and have been successfully applied in various domains.
Despite the effectiveness of GNNs, it is still challenging for GNNs to efficiently scale to large graphs.
As a remedy, distributed computing becomes a promising solution of training large-scale GNNs.
arXiv Detail & Related papers (2022-11-01T01:57:00Z) - A Comprehensive Study on Large-Scale Graph Training: Benchmarking and
Rethinking [124.21408098724551]
Large-scale graph training is a notoriously challenging problem for graph neural networks (GNNs)
We present a new ensembling training manner, named EnGCN, to address the existing issues.
Our proposed method has achieved new state-of-the-art (SOTA) performance on large-scale datasets.
arXiv Detail & Related papers (2022-10-14T03:43:05Z) - Hardware/Software Co-Programmable Framework for Computational SSDs to
Accelerate Deep Learning Service on Large-Scale Graphs [8.698995648930806]
Graph neural networks (GNNs) process large-scale graphs consisting of a hundred billion edges.
We propose a novel deep learning framework on large graphs, HolisticGNN, that provides an easy-to-use, near-storage inference infrastructure for fast, energy-efficient GNN processing.
arXiv Detail & Related papers (2022-01-23T06:08:18Z) - Training Graph Neural Networks with 1000 Layers [133.84813995275988]
We study reversible connections, group convolutions, weight tying, and equilibrium models to advance the memory and parameter efficiency of GNNs.
To the best of our knowledge, RevGNN-Deep is the deepest GNN in the literature by one order of magnitude.
arXiv Detail & Related papers (2021-06-14T15:03:00Z) - DistGNN: Scalable Distributed Training for Large-Scale Graph Neural
Networks [58.48833325238537]
Full-batch training on Graph Neural Networks (GNN) to learn the structure of large graphs is a critical problem that needs to scale to hundreds of compute nodes to be feasible.
In this paper, we presentGNN that optimize the well-known Deep Graph Library (DGL) for full-batch training on CPU clusters.
Our results on four common GNN benchmark datasets show up to 3.7x speed-up using a single CPU socket and up to 97x speed-up using 128 CPU sockets.
arXiv Detail & Related papers (2021-04-14T08:46:35Z) - BlockGNN: Towards Efficient GNN Acceleration Using Block-Circulant
Weight Matrices [9.406007544032848]
Graph Neural Networks (GNNs) are state-of-the-art algorithms for analyzing non-euclidean graph data.
How to inference GNNs in real time has become a challenging problem for some resource-limited edge-computing platforms.
We propose BlockGNN, a software- hardware co-design approach to realize efficient GNN acceleration.
arXiv Detail & Related papers (2021-04-13T14:09:22Z) - A Unified Lottery Ticket Hypothesis for Graph Neural Networks [82.31087406264437]
We present a unified GNN sparsification (UGS) framework that simultaneously prunes the graph adjacency matrix and the model weights.
We further generalize the popular lottery ticket hypothesis to GNNs for the first time, by defining a graph lottery ticket (GLT) as a pair of core sub-dataset and sparse sub-network.
arXiv Detail & Related papers (2021-02-12T21:52:43Z) - Eigen-GNN: A Graph Structure Preserving Plug-in for GNNs [95.63153473559865]
Graph Neural Networks (GNNs) are emerging machine learning models on graphs.
Most existing GNN models in practice are shallow and essentially feature-centric.
We show empirically and analytically that the existing shallow GNNs cannot preserve graph structures well.
We propose Eigen-GNN, a plug-in module to boost GNNs ability in preserving graph structures.
arXiv Detail & Related papers (2020-06-08T02:47:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.