ShiftAddNet: A Hardware-Inspired Deep Network
- URL: http://arxiv.org/abs/2010.12785v1
- Date: Sat, 24 Oct 2020 05:09:14 GMT
- Title: ShiftAddNet: A Hardware-Inspired Deep Network
- Authors: Haoran You, Xiaohan Chen, Yongan Zhang, Chaojian Li, Sicheng Li, Zihao
Liu, Zhangyang Wang, Yingyan Lin
- Abstract summary: ShiftAddNet is an energy-efficient multiplication-less deep neural network.
It leads to both energy-efficient inference and training, without compromising expressive capacity.
ShiftAddNet aggressively reduces over 80% hardware-quantified energy cost of DNNs training and inference, while offering comparable or better accuracies.
- Score: 87.18216601210763
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multiplication (e.g., convolution) is arguably a cornerstone of modern deep
neural networks (DNNs). However, intensive multiplications cause expensive
resource costs that challenge DNNs' deployment on resource-constrained edge
devices, driving several attempts for multiplication-less deep networks. This
paper presented ShiftAddNet, whose main inspiration is drawn from a common
practice in energy-efficient hardware implementation, that is, multiplication
can be instead performed with additions and logical bit-shifts. We leverage
this idea to explicitly parameterize deep networks in this way, yielding a new
type of deep network that involves only bit-shift and additive weight layers.
This hardware-inspired ShiftAddNet immediately leads to both energy-efficient
inference and training, without compromising the expressive capacity compared
to standard DNNs. The two complementary operation types (bit-shift and add)
additionally enable finer-grained control of the model's learning capacity,
leading to more flexible trade-off between accuracy and (training) efficiency,
as well as improved robustness to quantization and pruning. We conduct
extensive experiments and ablation studies, all backed up by our FPGA-based
ShiftAddNet implementation and energy measurements. Compared to existing DNNs
or other multiplication-less models, ShiftAddNet aggressively reduces over 80%
hardware-quantified energy cost of DNNs training and inference, while offering
comparable or better accuracies. Codes and pre-trained models are available at
https://github.com/RICE-EIC/ShiftAddNet.
Related papers
- Energy Efficient Hardware Acceleration of Neural Networks with
Power-of-Two Quantisation [0.0]
We show that a hardware neural network accelerator with PoT weights implemented on the Zynq UltraScale + MPSoC ZCU104 FPGA can be at least $1.4x$ more energy efficient than the uniform quantisation version.
arXiv Detail & Related papers (2022-09-30T06:33:40Z) - DenseShift: Towards Accurate and Efficient Low-Bit Power-of-Two
Quantization [27.231327287238102]
We propose the DenseShift network, which significantly improves the accuracy of Shift networks.
Our experiments on various computer vision and speech tasks demonstrate that DenseShift outperforms existing low-bit multiplication-free networks.
arXiv Detail & Related papers (2022-08-20T15:17:40Z) - DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware
Efficiency of Compact Neural Networks [29.46621102184345]
We propose a framework dubbed DepthShrinker to develop hardware-friendly compact networks.
Our framework delivers hardware-friendly compact networks that outperform both state-of-the-art efficient DNNs and compression techniques.
arXiv Detail & Related papers (2022-06-02T02:32:47Z) - ShiftAddNAS: Hardware-Inspired Search for More Accurate and Efficient
Neural Networks [42.28659737268829]
ShiftAddNAS can automatically search for more accurate and more efficient NNs.
ShiftAddNAS integrates the first hybrid search space that incorporates both multiplication-based and multiplication-free operators.
Experiments and ablation studies consistently validate the efficacy of ShiftAddNAS.
arXiv Detail & Related papers (2022-05-17T06:40:13Z) - DS-Net++: Dynamic Weight Slicing for Efficient Inference in CNNs and
Transformers [105.74546828182834]
We show a hardware-efficient dynamic inference regime, named dynamic weight slicing, which adaptively slice a part of network parameters for inputs with diverse difficulty levels.
We present dynamic slimmable network (DS-Net) and dynamic slice-able network (DS-Net++) by input-dependently adjusting filter numbers of CNNs and multiple dimensions in both CNNs and transformers.
arXiv Detail & Related papers (2021-09-21T09:57:21Z) - Adder Neural Networks [75.54239599016535]
We present adder networks (AdderNets) to trade massive multiplications in deep neural networks.
In AdderNets, we take the $ell_p$-norm distance between filters and input feature as the output response.
We show that the proposed AdderNets can achieve 75.7% Top-1 accuracy 92.3% Top-5 accuracy using ResNet-50 on the ImageNet dataset.
arXiv Detail & Related papers (2021-05-29T04:02:51Z) - Dynamic Slimmable Network [105.74546828182834]
We develop a dynamic network slimming regime named Dynamic Slimmable Network (DS-Net)
Our DS-Net is empowered with the ability of dynamic inference by the proposed double-headed dynamic gate.
It consistently outperforms its static counterparts as well as state-of-the-art static and dynamic model compression methods.
arXiv Detail & Related papers (2021-03-24T15:25:20Z) - AdderNet and its Minimalist Hardware Design for Energy-Efficient
Artificial Intelligence [111.09105910265154]
We present a novel minimalist hardware architecture using adder convolutional neural network (AdderNet)
The whole AdderNet can practically achieve 16% enhancement in speed.
We conclude the AdderNet is able to surpass all the other competitors.
arXiv Detail & Related papers (2021-01-25T11:31:52Z) - AdderNet: Do We Really Need Multiplications in Deep Learning? [159.174891462064]
We present adder networks (AdderNets) to trade massive multiplications in deep neural networks for much cheaper additions to reduce computation costs.
We develop a special back-propagation approach for AdderNets by investigating the full-precision gradient.
As a result, the proposed AdderNets can achieve 74.9% Top-1 accuracy 91.7% Top-5 accuracy using ResNet-50 on the ImageNet dataset.
arXiv Detail & Related papers (2019-12-31T06:56:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.