Hardware-Aware Graph Neural Network Automated Design for Edge Computing
Platforms
- URL: http://arxiv.org/abs/2303.10875v2
- Date: Thu, 13 Apr 2023 09:07:51 GMT
- Title: Hardware-Aware Graph Neural Network Automated Design for Edge Computing
Platforms
- Authors: Ao Zhou, Jianlei Yang, Yingjie Qi, Yumeng Shi, Tong Qiao, Weisheng
Zhao, Chunming Hu
- Abstract summary: HGNAS is proposed as the first Hardware-aware Graph Neural Architecture Search framework targeting resource constraint edge devices.
Results show that HGNAS can achieve about $10.6times$ speedup and $88.2%$ peak memory reduction with a negligible accuracy loss compared to DGCNN on various edge devices.
- Score: 9.345807588929734
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph neural networks (GNNs) have emerged as a popular strategy for handling
non-Euclidean data due to their state-of-the-art performance. However, most of
the current GNN model designs mainly focus on task accuracy, lacking in
considering hardware resources limitation and real-time requirements of edge
application scenarios. Comprehensive profiling of typical GNN models indicates
that their execution characteristics are significantly affected across
different computing platforms, which demands hardware awareness for efficient
GNN designs. In this work, HGNAS is proposed as the first Hardware-aware Graph
Neural Architecture Search framework targeting resource constraint edge
devices. By decoupling the GNN paradigm, HGNAS constructs a fine-grained design
space and leverages an efficient multi-stage search strategy to explore optimal
architectures within a few GPU hours. Moreover, HGNAS achieves hardware
awareness during the GNN architecture design by leveraging a hardware
performance predictor, which could balance the GNN model accuracy and
efficiency corresponding to the characteristics of targeted devices.
Experimental results show that HGNAS can achieve about $10.6\times$ speedup and
$88.2\%$ peak memory reduction with a negligible accuracy loss compared to
DGCNN on various edge devices, including Nvidia RTX3080, Jetson TX2, Intel
i7-8700K and Raspberry Pi 3B+.
Related papers
- HGNAS: Hardware-Aware Graph Neural Architecture Search for Edge Devices [11.1990060370675]
This work proposes a novel hardware-aware graph neural architecture search framework tailored for resource constraint edge devices, namely HGNAS.
HGNAS integrates an efficient GNN hardware performance predictor that evaluates the latency and peak memory usage of GNNs in milliseconds.
It can achieve up to a 10.6x speedup and an 82.5% peak memory reduction with negligible accuracy loss compared to DGCNN on ModelNet40.
arXiv Detail & Related papers (2024-08-23T05:11:22Z) - Quasar-ViT: Hardware-Oriented Quantization-Aware Architecture Search for Vision Transformers [56.37495946212932]
Vision transformers (ViTs) have demonstrated their superior accuracy for computer vision tasks compared to convolutional neural networks (CNNs)
This work proposes Quasar-ViT, a hardware-oriented quantization-aware architecture search framework for ViTs.
arXiv Detail & Related papers (2024-07-25T16:35:46Z) - Spiker+: a framework for the generation of efficient Spiking Neural
Networks FPGA accelerators for inference at the edge [49.42371633618761]
Spiker+ is a framework for generating efficient, low-power, and low-area customized Spiking Neural Networks (SNN) accelerators on FPGA for inference at the edge.
Spiker+ is tested on two benchmark datasets, the MNIST and the Spiking Heidelberg Digits (SHD)
arXiv Detail & Related papers (2024-01-02T10:42:42Z) - MaGNAS: A Mapping-Aware Graph Neural Architecture Search Framework for
Heterogeneous MPSoC Deployment [8.29394286023338]
We propose a novel unified design-mapping approach for efficient processing of vision GNN workloads on heterogeneous MPSoC platforms.
MaGNAS employs a two-tier evolutionary search to identify optimal GNNs and mapping pairings that yield the best performance trade-offs.
Our experimental results demonstrate that MaGNAS is able to provide 1.57x latency speedup and is 3.38x more energy-efficient for several vision datasets executed on the Xavier MPSoC vs. the GPU-only deployment.
arXiv Detail & Related papers (2023-07-16T14:56:50Z) - GHOST: A Graph Neural Network Accelerator using Silicon Photonics [4.226093500082746]
Graph neural networks (GNNs) have emerged as a powerful approach for modelling and learning from graph-structured data.
We present GHOST, the first silicon-photonic hardware accelerator for GNNs.
arXiv Detail & Related papers (2023-07-04T15:37:20Z) - A Comprehensive Study on Large-Scale Graph Training: Benchmarking and
Rethinking [124.21408098724551]
Large-scale graph training is a notoriously challenging problem for graph neural networks (GNNs)
We present a new ensembling training manner, named EnGCN, to address the existing issues.
Our proposed method has achieved new state-of-the-art (SOTA) performance on large-scale datasets.
arXiv Detail & Related papers (2022-10-14T03:43:05Z) - Hardware/Software Co-Programmable Framework for Computational SSDs to
Accelerate Deep Learning Service on Large-Scale Graphs [8.698995648930806]
Graph neural networks (GNNs) process large-scale graphs consisting of a hundred billion edges.
We propose a novel deep learning framework on large graphs, HolisticGNN, that provides an easy-to-use, near-storage inference infrastructure for fast, energy-efficient GNN processing.
arXiv Detail & Related papers (2022-01-23T06:08:18Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - BlockGNN: Towards Efficient GNN Acceleration Using Block-Circulant
Weight Matrices [9.406007544032848]
Graph Neural Networks (GNNs) are state-of-the-art algorithms for analyzing non-euclidean graph data.
How to inference GNNs in real time has become a challenging problem for some resource-limited edge-computing platforms.
We propose BlockGNN, a software- hardware co-design approach to realize efficient GNN acceleration.
arXiv Detail & Related papers (2021-04-13T14:09:22Z) - A Unified Lottery Ticket Hypothesis for Graph Neural Networks [82.31087406264437]
We present a unified GNN sparsification (UGS) framework that simultaneously prunes the graph adjacency matrix and the model weights.
We further generalize the popular lottery ticket hypothesis to GNNs for the first time, by defining a graph lottery ticket (GLT) as a pair of core sub-dataset and sparse sub-network.
arXiv Detail & Related papers (2021-02-12T21:52:43Z) - PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with
Pattern-based Weight Pruning [57.20262984116752]
We introduce a new dimension, fine-grained pruning patterns inside the coarse-grained structures, revealing a previously unknown point in design space.
With the higher accuracy enabled by fine-grained pruning patterns, the unique insight is to use the compiler to re-gain and guarantee high hardware efficiency.
arXiv Detail & Related papers (2020-01-01T04:52:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.