Winograd Algorithm for AdderNet
- URL: http://arxiv.org/abs/2105.05530v1
- Date: Wed, 12 May 2021 09:13:34 GMT
- Title: Winograd Algorithm for AdderNet
- Authors: Wenshuo Li, Hanting Chen, Mingqiang Huang, Xinghao Chen, Chunjing Xu,
Yunhe Wang
- Abstract summary: Adder neural network (AdderNet) is a new kind of deep model that replaces the original massive multiplications in convolutions by additions.
This paper studies the winograd algorithm, which is a widely used fast algorithm for accelerating convolution and saving the computational costs.
- Score: 54.93995545896655
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Adder neural network (AdderNet) is a new kind of deep model that replaces the
original massive multiplications in convolutions by additions while preserving
the high performance. Since the hardware complexity of additions is much lower
than that of multiplications, the overall energy consumption is thus reduced
significantly. To further optimize the hardware overhead of using AdderNet,
this paper studies the winograd algorithm, which is a widely used fast
algorithm for accelerating convolution and saving the computational costs.
Unfortunately, the conventional Winograd algorithm cannot be directly applied
to AdderNets since the distributive law in multiplication is not valid for the
l1-norm. Therefore, we replace the element-wise multiplication in the Winograd
equation by additions and then develop a new set of transform matrixes that can
enhance the representation ability of output features to maintain the
performance. Moreover, we propose the l2-to-l1 training strategy to mitigate
the negative impacts caused by formal inconsistency. Experimental results on
both FPGA and benchmarks show that the new method can further reduce the energy
consumption without affecting the accuracy of the original AdderNet.
Related papers
- ShiftAddAug: Augment Multiplication-Free Tiny Neural Network with Hybrid Computation [7.143509813507735]
ShiftAddAug uses costly multiplication to augment efficient but less powerful multiplication-free operators, improving performance without any inference overhead.
It secures up to a 4.95% increase in accuracy on the CIFAR100 compared to its directly trained counterparts, even surpassing the performance of multiplicative NNs.
arXiv Detail & Related papers (2024-07-03T07:56:51Z) - Adder Neural Networks [75.54239599016535]
We present adder networks (AdderNets) to trade massive multiplications in deep neural networks.
In AdderNets, we take the $ell_p$-norm distance between filters and input feature as the output response.
We show that the proposed AdderNets can achieve 75.7% Top-1 accuracy 92.3% Top-5 accuracy using ResNet-50 on the ImageNet dataset.
arXiv Detail & Related papers (2021-05-29T04:02:51Z) - AdderNet and its Minimalist Hardware Design for Energy-Efficient
Artificial Intelligence [111.09105910265154]
We present a novel minimalist hardware architecture using adder convolutional neural network (AdderNet)
The whole AdderNet can practically achieve 16% enhancement in speed.
We conclude the AdderNet is able to surpass all the other competitors.
arXiv Detail & Related papers (2021-01-25T11:31:52Z) - ShiftAddNet: A Hardware-Inspired Deep Network [87.18216601210763]
ShiftAddNet is an energy-efficient multiplication-less deep neural network.
It leads to both energy-efficient inference and training, without compromising expressive capacity.
ShiftAddNet aggressively reduces over 80% hardware-quantified energy cost of DNNs training and inference, while offering comparable or better accuracies.
arXiv Detail & Related papers (2020-10-24T05:09:14Z) - AdderSR: Towards Energy Efficient Image Super-Resolution [127.61437479490047]
This paper studies the single image super-resolution problem using adder neural networks (AdderNet)
Compared with convolutional neural networks, AdderNet utilizing additions to calculate the output features thus avoid massive energy consumptions of conventional multiplications.
arXiv Detail & Related papers (2020-09-18T15:29:13Z) - LANCE: Efficient Low-Precision Quantized Winograd Convolution for Neural
Networks Based on Graphics Processing Units [6.110973485878557]
We propose an efficient low-precision quantized Winograd convolution algorithm, called LANCE, which combines the advantages of fast convolution and quantization techniques.
We show that our 8-bit quantized Winograd convolution improves the performance by up to 2.40x over the full-precision convolution with trivial accuracy loss.
arXiv Detail & Related papers (2020-03-19T09:46:50Z) - AdderNet: Do We Really Need Multiplications in Deep Learning? [159.174891462064]
We present adder networks (AdderNets) to trade massive multiplications in deep neural networks for much cheaper additions to reduce computation costs.
We develop a special back-propagation approach for AdderNets by investigating the full-precision gradient.
As a result, the proposed AdderNets can achieve 74.9% Top-1 accuracy 91.7% Top-5 accuracy using ResNet-50 on the ImageNet dataset.
arXiv Detail & Related papers (2019-12-31T06:56:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.