SGCN: Exploiting Compressed-Sparse Features in Deep Graph Convolutional
Network Accelerators
- URL: http://arxiv.org/abs/2301.10388v1
- Date: Wed, 25 Jan 2023 02:34:01 GMT
- Title: SGCN: Exploiting Compressed-Sparse Features in Deep Graph Convolutional
Network Accelerators
- Authors: Mingi Yoo, Jaeyong Song, Jounghoo Lee, Namhyung Kim, Youngsok Kim, and
Jinho Lee
- Abstract summary: Graph convolutional networks (GCNs) are becoming increasingly popular as they overcome the limited applicability of prior neural networks.
In this paper, we propose SGCN, a fast and energy-efficient GCN accelerator.
We show that SGCN achieves 1.71x speedup and 43.9% higher energy efficiency compared to the existing accelerators.
- Score: 6.582242235154822
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Graph convolutional networks (GCNs) are becoming increasingly popular as they
overcome the limited applicability of prior neural networks. A GCN takes as
input an arbitrarily structured graph and executes a series of layers which
exploit the graph's structure to calculate their output features. One recent
trend in GCNs is the use of deep network architectures. As opposed to the
traditional GCNs which only span around two to five layers deep, modern GCNs
now incorporate tens to hundreds of layers with the help of residual
connections. From such deep GCNs, we find an important characteristic that they
exhibit very high intermediate feature sparsity. We observe that with deep
layers and residual connections, the number of zeros in the intermediate
features sharply increases. This reveals a new opportunity for accelerators to
exploit in GCN executions that was previously not present.
In this paper, we propose SGCN, a fast and energy-efficient GCN accelerator
which fully exploits the sparse intermediate features of modern GCNs. SGCN
suggests several techniques to achieve significantly higher performance and
energy efficiency than the existing accelerators. First, SGCN employs a
GCN-friendly feature compression format. We focus on reducing the off-chip
memory traffic, which often is the bottleneck for GCN executions. Second, we
propose microarchitectures for seamlessly handling the compressed feature
format. Third, to better handle locality in the existence of the varying
sparsity, SGCN employs sparsity-aware cooperation. Sparsity-aware cooperation
creates a pattern that exhibits multiple reuse windows, such that the cache can
capture diverse sizes of working sets and therefore adapt to the varying level
of sparsity. We show that SGCN achieves 1.71x speedup and 43.9% higher energy
efficiency compared to the existing accelerators.
Related papers
- GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for
Memory-Efficient Graph Convolutional Neural Networks [4.669338722185048]
A unique property of Graph convolutional neural networks (GCNs) is that its two primary execution stages, aggregation and combination, exhibit drastically different dataflows.
We present GROW, a GCN accelerator based on Gustavson's algorithm to architect a row-wise product based sparse-dense GEMM accelerator.
arXiv Detail & Related papers (2022-03-01T00:26:31Z) - GCoD: Graph Convolutional Network Acceleration via Dedicated Algorithm
and Accelerator Co-Design [27.311994997480745]
Graph Convolutional Networks (GCNs) have emerged as the state-of-the-art graph learning model.
It can be notoriously challenging to inference GCNs over large graph datasets.
This paper proposes a GCN algorithm and accelerator Co-Design framework dubbed GCoD which can largely alleviate the aforementioned GCN irregularity.
arXiv Detail & Related papers (2021-12-22T00:30:50Z) - Multi-scale Graph Convolutional Networks with Self-Attention [2.66512000865131]
Graph convolutional networks (GCNs) have achieved remarkable learning ability for dealing with various graph structural data.
Over-smoothing phenomenon as a crucial issue of GCNs remains to be solved and investigated.
We propose two novel multi-scale GCN frameworks by incorporating self-attention mechanism and multi-scale information into the design of GCNs.
arXiv Detail & Related papers (2021-12-04T04:41:24Z) - LW-GCN: A Lightweight FPGA-based Graph Convolutional Network Accelerator [14.145707219377917]
Graph convolutional networks (GCNs) have been introduced to effectively process non-euclidean graph data.
LW-GCN decomposes the main GCN operations into sparse-dense matrix multiplication (SDMM) and dense matrix multiplication (DMM)
Compared to existing CPU, GPU, and state-of-the-art FPGA-based accelerator, LW-GCN reduces latency by up to 60x, 12x and 1.7x and increases power efficiency by up to 912x, 511x and 3.87x, respectively.
arXiv Detail & Related papers (2021-11-04T22:29:53Z) - Towards Efficient Graph Convolutional Networks for Point Cloud Handling [181.59146413326056]
We aim at improving the computational efficiency of graph convolutional networks (GCNs) for learning on point clouds.
A series of experiments show that optimized networks have reduced computational complexity, decreased memory consumption, and accelerated inference speed.
arXiv Detail & Related papers (2021-04-12T17:59:16Z) - Tackling Over-Smoothing for General Graph Convolutional Networks [88.71154017107257]
We study how general GCNs act with the increase in depth, including generic GCN, GCN with bias, ResGCN, and APPNP.
We propose DropEdge to alleviate over-smoothing by randomly removing a certain number of edges at each training epoch.
arXiv Detail & Related papers (2020-08-22T16:14:01Z) - DeeperGCN: All You Need to Train Deeper GCNs [66.64739331859226]
Graph Convolutional Networks (GCNs) have been drawing significant attention with the power of representation learning on graphs.
Unlike Convolutional Neural Networks (CNNs), which are able to take advantage of stacking very deep layers, GCNs suffer from vanishing gradient, over-smoothing and over-fitting issues when going deeper.
This paper proposes DeeperGCN that is capable of successfully and reliably training very deep GCNs.
arXiv Detail & Related papers (2020-06-13T23:00:22Z) - Graph Highway Networks [77.38665506495553]
Graph Convolution Networks (GCN) are widely used in learning graph representations due to their effectiveness and efficiency.
They suffer from the notorious over-smoothing problem, in which the learned representations converge to alike vectors when many layers are stacked.
We propose Graph Highway Networks (GHNet) which utilize gating units to balance the trade-off between homogeneity and heterogeneity in the GCN learning process.
arXiv Detail & Related papers (2020-04-09T16:26:43Z) - L$^2$-GCN: Layer-Wise and Learned Efficient Training of Graph
Convolutional Networks [118.37805042816784]
Graph convolution networks (GCN) are increasingly popular in many applications, yet remain notoriously hard to train over large graph datasets.
We propose a novel efficient layer-wise training framework for GCN (L-GCN), that disentangles feature aggregation and feature transformation during training.
Experiments show that L-GCN is faster than state-of-the-arts by at least an order of magnitude, with a consistent of memory usage not dependent on dataset size.
arXiv Detail & Related papers (2020-03-30T16:37:56Z) - LightGCN: Simplifying and Powering Graph Convolution Network for
Recommendation [100.76229017056181]
Graph Convolution Network (GCN) has become new state-of-the-art for collaborative filtering.
In this work, we aim to simplify the design of GCN to make it more concise and appropriate for recommendation.
We propose a new model named LightGCN, including only the most essential component in GCN -- neighborhood aggregation.
arXiv Detail & Related papers (2020-02-06T06:53:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.