ReActNet: Towards Precise Binary Neural Network with Generalized
Activation Functions
- URL: http://arxiv.org/abs/2003.03488v2
- Date: Mon, 13 Jul 2020 03:05:51 GMT
- Title: ReActNet: Towards Precise Binary Neural Network with Generalized
Activation Functions
- Authors: Zechun Liu and Zhiqiang Shen and Marios Savvides and Kwang-Ting Cheng
- Abstract summary: We propose several ideas for enhancing a binary network to close its accuracy gap from real-valued networks without incurring any additional computational cost.
We first construct a baseline network by modifying and binarizing a compact real-valued network with parameter-free shortcuts.
We show that the proposed ReActNet outperforms all the state-of-the-arts by a large margin.
- Score: 76.05981545084738
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose several ideas for enhancing a binary network to
close its accuracy gap from real-valued networks without incurring any
additional computational cost. We first construct a baseline network by
modifying and binarizing a compact real-valued network with parameter-free
shortcuts, bypassing all the intermediate convolutional layers including the
downsampling layers. This baseline network strikes a good trade-off between
accuracy and efficiency, achieving superior performance than most of existing
binary networks at approximately half of the computational cost. Through
extensive experiments and analysis, we observed that the performance of binary
networks is sensitive to activation distribution variations. Based on this
important observation, we propose to generalize the traditional Sign and PReLU
functions, denoted as RSign and RPReLU for the respective generalized
functions, to enable explicit learning of the distribution reshape and shift at
near-zero extra cost. Lastly, we adopt a distributional loss to further enforce
the binary network to learn similar output distributions as those of a
real-valued network. We show that after incorporating all these ideas, the
proposed ReActNet outperforms all the state-of-the-arts by a large margin.
Specifically, it outperforms Real-to-Binary Net and MeliusNet29 by 4.0% and
3.6% respectively for the top-1 accuracy and also reduces the gap to its
real-valued counterpart to within 3.0% top-1 accuracy on ImageNet dataset. Code
and models are available at: https://github.com/liuzechun/ReActNet.
Related papers
- BiHRNet: A Binary high-resolution network for Human Pose Estimation [11.250422970707415]
We propose a binary human pose estimator named BiHRNet, whose weights and activations are expressed as $pm$1.
BiHRNet retains the keypoint extraction ability of HRNet, while using fewer computing resources by adapting binary neural network (BNN)
We show BiHRNet achieves a PCKh of 87.9 on the MPII dataset, which outperforms all binary pose estimation networks.
arXiv Detail & Related papers (2023-11-17T03:01:37Z) - BiFSMNv2: Pushing Binary Neural Networks for Keyword Spotting to
Real-Network Performance [54.214426436283134]
Deep neural networks, such as the Deep-FSMN, have been widely studied for keyword spotting (KWS) applications.
We present a strong yet efficient binary neural network for KWS, namely BiFSMNv2, pushing it to the real-network accuracy performance.
We highlight that benefiting from the compact architecture and optimized hardware kernel, BiFSMNv2 can achieve an impressive 25.1x speedup and 20.2x storage-saving on edge hardware.
arXiv Detail & Related papers (2022-11-13T18:31:45Z) - BiFSMN: Binary Neural Network for Keyword Spotting [47.46397208920726]
BiFSMN is an accurate and extreme-efficient binary neural network for KWS.
We show that BiFSMN can achieve an impressive 22.3x speedup and 15.5x storage-saving on real-world edge hardware.
arXiv Detail & Related papers (2022-02-14T05:16:53Z) - Distribution-sensitive Information Retention for Accurate Binary Neural
Network [49.971345958676196]
We present a novel Distribution-sensitive Information Retention Network (DIR-Net) to retain the information of the forward activations and backward gradients.
Our DIR-Net consistently outperforms the SOTA binarization approaches under mainstream and compact architectures.
We conduct our DIR-Net on real-world resource-limited devices which achieves 11.1 times storage saving and 5.4 times speedup.
arXiv Detail & Related papers (2021-09-25T10:59:39Z) - Enabling certification of verification-agnostic networks via
memory-efficient semidefinite programming [97.40955121478716]
We propose a first-order dual SDP algorithm that requires memory only linear in the total number of network activations.
We significantly improve L-inf verified robust accuracy from 1% to 88% and 6% to 40% respectively.
We also demonstrate tight verification of a quadratic stability specification for the decoder of a variational autoencoder.
arXiv Detail & Related papers (2020-10-22T12:32:29Z) - High-Capacity Expert Binary Networks [56.87581500474093]
Network binarization is a promising hardware-aware direction for creating efficient deep models.
Despite its memory and computational advantages, reducing the accuracy gap between binary models and their real-valued counterparts remains an unsolved challenging research problem.
We propose Expert Binary Convolution, which, for the first time, tailors conditional computing to binary networks by learning to select one data-specific expert binary filter at a time conditioned on input features.
arXiv Detail & Related papers (2020-10-07T17:58:10Z) - Towards Lossless Binary Convolutional Neural Networks Using Piecewise
Approximation [4.023728681102073]
CNNs can significantly reduce the number of arithmetic operations and the size of memory storage.
However, the accuracy degradation of single and multiple binary CNNs is unacceptable for modern architectures.
We propose a Piecewise Approximation scheme for multiple binary CNNs which lessens accuracy loss by approximating full precision weights and activations.
arXiv Detail & Related papers (2020-08-08T13:32:33Z) - Training Binary Neural Networks with Real-to-Binary Convolutions [52.91164959767517]
We show how to train binary networks to within a few percent points of the full precision counterpart.
We show how to build a strong baseline, which already achieves state-of-the-art accuracy.
We show that, when putting all of our improvements together, the proposed model beats the current state of the art by more than 5% top-1 accuracy on ImageNet.
arXiv Detail & Related papers (2020-03-25T17:54:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.