A scalable and efficient convolutional neural network accelerator using
HLS for a System on Chip design
- URL: http://arxiv.org/abs/2004.13075v2
- Date: Wed, 7 Oct 2020 06:58:55 GMT
- Title: A scalable and efficient convolutional neural network accelerator using
HLS for a System on Chip design
- Authors: Kim Bjerge, Jonathan Horsted Schougaard and Daniel Ejnar Larsen
- Abstract summary: The presented CNNA has a scalable architecture which uses High Level Synthesis (HLS) and SystemC for the hardware accelerator.
It is able to accelerate any Convolutional Neural Network exported from Python and supports a combination of convolutional, max-pooling, and fully connected layers.
It was able to perform inference in 2.0 seconds, while having an average power consumption of 2.63 W, which corresponds to a power efficiency of 6.0 GOPS/W.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a configurable Convolutional Neural Network Accelerator
(CNNA) for a System on Chip design (SoC). The goal was to accelerate inference
of different deep learning networks on an embedded SoC platform. The presented
CNNA has a scalable architecture which uses High Level Synthesis (HLS) and
SystemC for the hardware accelerator. It is able to accelerate any
Convolutional Neural Network (CNN) exported from Python and supports a
combination of convolutional, max-pooling, and fully connected layers. A
training method with fixed-point quantized weights is proposed and presented in
the paper. The CNNA is template-based, enabling it to scale for different
targets of the Xilinx Zynq platform. This approach enables design space
exploration, which makes it possible to explore several configurations of the
CNNA during C- and RTL-simulation, fitting it to the desired platform and
model. The CNN VGG16 was used to test the solution on a Xilinx Ultra96 board
using PYNQ. The result gave a high level of accuracy in training with an
auto-scaled fixed-point Q2.14 format compared to a similar floating-point
model. It was able to perform inference in 2.0 seconds, while having an average
power consumption of 2.63 W, which corresponds to a power efficiency of 6.0
GOPS/W.
Related papers
- TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - Kronecker-Factored Approximate Curvature for Modern Neural Network
Architectures [85.76673783330334]
Two different settings of linear weight-sharing layers motivate two flavours of Kronecker-Factored Approximate Curvature (K-FAC)
We show they are exact for deep linear networks with weight-sharing in their respective setting.
We observe little difference between these two K-FAC variations when using them to train both a graph neural network and a vision transformer.
arXiv Detail & Related papers (2023-11-01T16:37:00Z) - LPYOLO: Low Precision YOLO for Face Detection on FPGA [1.7188280334580197]
Face detection on surveillance systems is the most expected application on the security market.
TinyYolov3 architecture is redesigned and deployed for face detection.
Model is converted to an HLS based application with using FINN framework and FINN-HLS library.
arXiv Detail & Related papers (2022-07-21T13:54:52Z) - Towards Enabling Dynamic Convolution Neural Network Inference for Edge
Intelligence [0.0]
Recent advances in edge intelligence require CNN inference on edge network to increase throughput and reduce latency.
To provide flexibility, dynamic parameter allocation to different mobile devices is required to implement either a predefined or defined on-the-fly CNN architecture.
We propose a library-based approach to design scalable and dynamic distributed CNN inference on the fly.
arXiv Detail & Related papers (2022-02-18T22:33:42Z) - An Adaptive Device-Edge Co-Inference Framework Based on Soft
Actor-Critic [72.35307086274912]
High-dimension parameter model and large-scale mathematical calculation restrict execution efficiency, especially for Internet of Things (IoT) devices.
We propose a new Deep Reinforcement Learning (DRL)-Soft Actor Critic for discrete (SAC-d), which generates the emphexit point, emphexit point, and emphcompressing bits by soft policy iterations.
Based on the latency and accuracy aware reward design, such an computation can well adapt to the complex environment like dynamic wireless channel and arbitrary processing, and is capable of supporting the 5G URL
arXiv Detail & Related papers (2022-01-09T09:31:50Z) - A quantum algorithm for training wide and deep classical neural networks [72.2614468437919]
We show that conditions amenable to classical trainability via gradient descent coincide with those necessary for efficiently solving quantum linear systems.
We numerically demonstrate that the MNIST image dataset satisfies such conditions.
We provide empirical evidence for $O(log n)$ training of a convolutional neural network with pooling.
arXiv Detail & Related papers (2021-07-19T23:41:03Z) - Quantized Neural Networks via {-1, +1} Encoding Decomposition and
Acceleration [83.84684675841167]
We propose a novel encoding scheme using -1, +1 to decompose quantized neural networks (QNNs) into multi-branch binary networks.
We validate the effectiveness of our method on large-scale image classification, object detection, and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-18T03:11:15Z) - ANNETTE: Accurate Neural Network Execution Time Estimation with Stacked
Models [56.21470608621633]
We propose a time estimation framework to decouple the architectural search from the target hardware.
The proposed methodology extracts a set of models from micro- kernel and multi-layer benchmarks and generates a stacked model for mapping and network execution time estimation.
We compare estimation accuracy and fidelity of the generated mixed models, statistical models with the roofline model, and a refined roofline model for evaluation.
arXiv Detail & Related papers (2021-05-07T11:39:05Z) - Exploration of Hardware Acceleration Methods for an XNOR Traffic Signs
Classifier [0.0]
In this work, we explore the possibility of accelerating XNOR networks for traffic sign classification.
We propose a custom HDL accelerator for XNOR networks, which enables the inference with almost 450 fps.
Even better results are obtained with the second method - the Xilinx FINN accelerator - enabling to process input images with around 550 frame rate.
arXiv Detail & Related papers (2021-04-06T06:01:57Z) - A Tutorial on Quantum Convolutional Neural Networks (QCNN) [11.79760591464748]
Convolutional Neural Network (CNN) is a popular model in computer vision.
CNN is challenging to learn efficiently if the given dimension of data or model becomes too large.
Quantum Convolutional Neural Network (QCNN) provides a new solution to a problem to solve with CNN using a quantum computing environment.
arXiv Detail & Related papers (2020-09-20T12:29:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.