Benchmarking Quantized Neural Networks on FPGAs with FINN
- URL: http://arxiv.org/abs/2102.01341v1
- Date: Tue, 2 Feb 2021 06:42:07 GMT
- Title: Benchmarking Quantized Neural Networks on FPGAs with FINN
- Authors: Quentin Ducasse, Pascal Cotret, Lo\"ic Lagadec, Robert Stewart
- Abstract summary: Using lower precision comes at the cost of negligible loss in accuracy.
This article aims to assess the impact of mixed-precision when applied to neural networks deployed on FPGAs.
- Score: 0.42439262432068253
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The ever-growing cost of both training and inference for state-of-the-art
neural networks has brought literature to look upon ways to cut off resources
used with a minimal impact on accuracy. Using lower precision comes at the cost
of negligible loss in accuracy. While training neural networks may require a
powerful setup, deploying a network must be possible on low-power and
low-resource hardware architectures. Reconfigurable architectures have proven
to be more powerful and flexible than GPUs when looking at a specific
application. This article aims to assess the impact of mixed-precision when
applied to neural networks deployed on FPGAs. While several frameworks exist
that create tools to deploy neural networks using reduced-precision, few of
them assess the importance of quantization and the framework quality. FINN and
Brevitas, two frameworks from Xilinx labs, are used to assess the impact of
quantization on neural networks using 2 to 8 bit precisions and weights with
several parallelization configurations. Equivalent accuracy can be obtained
using lower-precision representation and enough training. However, the
compressed network can be better parallelized allowing the deployed network
throughput to be 62 times faster. The benchmark set up in this work is
available in a public repository (https://github.com/QDucasse/nn benchmark).
Related papers
Err
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.