GNN-DSE: Automated Accelerator Optimization Aided by Graph Neural
Networks
- URL: http://arxiv.org/abs/2111.08848v1
- Date: Wed, 17 Nov 2021 00:36:08 GMT
- Title: GNN-DSE: Automated Accelerator Optimization Aided by Graph Neural
Networks
- Authors: Atefeh Sohrabizadeh, Yunsheng Bai, Yizhou Sun, and Jason Cong
- Abstract summary: High-level synthesis (HLS) has freed the computer architects from developing their designs in a very low-level language.
We propose to solve this problem by modeling the HLS tool with a graph neural network (GNN) that is trained to be used for a wide range of applications.
- Score: 31.11645988046767
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-level synthesis (HLS) has freed the computer architects from developing
their designs in a very low-level language and needing to exactly specify how
the data should be transferred in register-level. With the help of HLS, the
hardware designers must describe only a high-level behavioral flow of the
design. Despite this, it still can take weeks to develop a high-performance
architecture mainly because there are many design choices at a higher level
that requires more time to explore. It also takes several minutes to hours to
get feedback from the HLS tool on the quality of each design candidate. In this
paper, we propose to solve this problem by modeling the HLS tool with a graph
neural network (GNN) that is trained to be used for a wide range of
applications. The experimental results demonstrate that by employing the
GNN-based model, we are able to estimate the quality of design in milliseconds
with high accuracy which can help us search through the solution space very
quickly.
Related papers
- Intelligent4DSE: Optimizing High-Level Synthesis Design Space Exploration with Graph Neural Networks and Large Language Models [3.8429489584622156]
We propose CoGNNs-LLMEA, a framework that integrates a graph neural network with task-adaptive message passing and a large language model-enhanced evolutionary algorithm.
As a predictive model, CoGNNs directly leverages intermediate representations generated from source code after compiler front-end processing, enabling prediction of quality of results (QoR) without invoking HLS tools.
CoGNNs achieves state-of-the-art prediction accuracy in post-HLS QoR prediction, reducing mean prediction errors by 2.8$times$ for latency and 3.4$times$ for resource utilization compared to baseline models
arXiv Detail & Related papers (2025-04-28T10:08:56Z) - Neural Architecture Codesign for Fast Physics Applications [0.8692847090818803]
We develop a pipeline to streamline neural architecture codesign for physics applications.
We employ neural architecture search and network compression in a two-stage approach to discover hardware efficient models.
arXiv Detail & Related papers (2025-01-09T19:00:03Z) - Computation-friendly Graph Neural Network Design by Accumulating Knowledge on Large Language Models [20.31388126105889]
Graph Neural Networks (GNNs) have shown remarkable success but are hampered by the complexity of their architecture designs.
To reduce human workload, researchers try to develop automated algorithms to design GNNs.
arXiv Detail & Related papers (2024-08-13T08:22:01Z) - NNsight and NDIF: Democratizing Access to Open-Weight Foundation Model Internals [58.83169560132308]
We introduce NNsight and NDIF, technologies that work in tandem to enable scientific study of the representations and computations learned by very large neural networks.
arXiv Detail & Related papers (2024-07-18T17:59:01Z) - Cross-Modality Program Representation Learning for Electronic Design Automation with High-Level Synthesis [45.471039079664656]
Domain-specific accelerators (DSAs) have gained popularity for applications such as deep learning and autonomous driving.
We propose ProgSG, a model that allows interaction between the source code sequence modality and the graph modality in a deep and fine-grained way.
We show that ProgSG reduces the RMSE of design performance predictions by up to $22%$, and identifies designs with an average of $1.10times$.
arXiv Detail & Related papers (2024-06-13T22:34:58Z) - AutoHLS: Learning to Accelerate Design Space Exploration for HLS Designs [10.690389829735661]
This paper proposes a novel framework called AutoHLS, which integrates a deep neural network (DNN) with Bayesian optimization (BO) to accelerate HLS hardware design optimization.
Our experimental results demonstrate up to a 70-fold speedup in exploration time.
arXiv Detail & Related papers (2024-03-15T21:14:44Z) - On the Viability of using LLMs for SW/HW Co-Design: An Example in
Designing CiM DNN Accelerators [14.02304927398616]
Deep Neural Networks (DNNs) have demonstrated impressive performance across a wide range of tasks.
deploying DNNs on edge devices poses significant challenges due to stringent power and computational budgets.
We present a novel approach that leverages Large Language Models (LLMs) to address this issue.
arXiv Detail & Related papers (2023-06-12T07:50:53Z) - Fast GraspNeXt: A Fast Self-Attention Neural Network Architecture for
Multi-task Learning in Computer Vision Tasks for Robotic Grasping on the Edge [80.88063189896718]
High architectural and computational complexity can result in poor suitability for deployment on embedded devices.
Fast GraspNeXt is a fast self-attention neural network architecture tailored for embedded multi-task learning in computer vision tasks for robotic grasping.
arXiv Detail & Related papers (2023-04-21T18:07:14Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - A Graph Deep Learning Framework for High-Level Synthesis Design Space
Exploration [11.154086943903696]
High-Level Synthesis is a solution for fast prototyping application-specific hardware.
We propose HLS, for the first time in the literature, graph neural networks that jointly predict acceleration performance and hardware costs.
We show that our approach achieves prediction accuracy comparable with that of commonly used simulators.
arXiv Detail & Related papers (2021-11-29T18:17:45Z) - Design Space for Graph Neural Networks [81.88707703106232]
We study the architectural design space for Graph Neural Networks (GNNs) which consists of 315,000 different designs over 32 different predictive tasks.
Our key results include: (1) A comprehensive set of guidelines for designing well-performing GNNs; (2) while best GNN designs for different tasks vary significantly, the GNN task space allows for transferring the best designs across different tasks; (3) models discovered using our design space achieve state-of-the-art performance.
arXiv Detail & Related papers (2020-11-17T18:59:27Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - NAS-Count: Counting-by-Density with Neural Architecture Search [74.92941571724525]
We automate the design of counting models with Neural Architecture Search (NAS)
We introduce an end-to-end searched encoder-decoder architecture, Automatic Multi-Scale Network (AMSNet)
arXiv Detail & Related papers (2020-02-29T09:18:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.