Join the High Accuracy Club on ImageNet with A Binary Neural Network
Ticket
- URL: http://arxiv.org/abs/2211.12933v1
- Date: Wed, 23 Nov 2022 13:08:58 GMT
- Title: Join the High Accuracy Club on ImageNet with A Binary Neural Network
Ticket
- Authors: Nianhui Guo, Joseph Bethge, Christoph Meinel, Haojin Yang
- Abstract summary: We focus on a problem: how can a binary neural network achieve the crucial accuracy level (e.g., 80%) on ILSVRC-2012 ImageNet?
We design a novel binary architecture BNext based on a comprehensive study of binary architectures and their optimization process.
We propose a novel knowledge-distillation technique to alleviate the counter-intuitive overfitting problem observed when attempting to train extremely accurate binary models.
- Score: 10.552465253379134
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Binary neural networks are the extreme case of network quantization, which
has long been thought of as a potential edge machine learning solution.
However, the significant accuracy gap to the full-precision counterparts
restricts their creative potential for mobile applications. In this work, we
revisit the potential of binary neural networks and focus on a compelling but
unanswered problem: how can a binary neural network achieve the crucial
accuracy level (e.g., 80%) on ILSVRC-2012 ImageNet? We achieve this goal by
enhancing the optimization process from three complementary perspectives: (1)
We design a novel binary architecture BNext based on a comprehensive study of
binary architectures and their optimization process. (2) We propose a novel
knowledge-distillation technique to alleviate the counter-intuitive overfitting
problem observed when attempting to train extremely accurate binary models. (3)
We analyze the data augmentation pipeline for binary networks and modernize it
with up-to-date techniques from full-precision models. The evaluation results
on ImageNet show that BNext, for the first time, pushes the binary model
accuracy boundary to 80.57% and significantly outperforms all the existing
binary networks. Code and trained models are available at: (blind URL, see
appendix).
Related papers
- Towards Accurate Binary Neural Networks via Modeling Contextual
Dependencies [52.691032025163175]
Existing Binary Neural Networks (BNNs) operate mainly on local convolutions with binarization function.
We present new designs of binary neural modules, which enables leading binary neural modules by a large margin.
arXiv Detail & Related papers (2022-09-03T11:51:04Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - Improving Accuracy of Binary Neural Networks using Unbalanced Activation
Distribution [12.46127622357824]
We show that unbalanced activation distribution can actually improve the accuracy of BNNs.
We also show that adjusting the threshold values of binary activation functions results in the unbalanced distribution of the binary activation.
Experimental results show that the accuracy of previous BNN models can be improved by simply shifting the threshold values of binary activation functions.
arXiv Detail & Related papers (2020-12-02T02:49:53Z) - High-Capacity Expert Binary Networks [56.87581500474093]
Network binarization is a promising hardware-aware direction for creating efficient deep models.
Despite its memory and computational advantages, reducing the accuracy gap between binary models and their real-valued counterparts remains an unsolved challenging research problem.
We propose Expert Binary Convolution, which, for the first time, tailors conditional computing to binary networks by learning to select one data-specific expert binary filter at a time conditioned on input features.
arXiv Detail & Related papers (2020-10-07T17:58:10Z) - Binarizing MobileNet via Evolution-based Searching [66.94247681870125]
We propose a use of evolutionary search to facilitate the construction and training scheme when binarizing MobileNet.
Inspired by one-shot architecture search frameworks, we manipulate the idea of group convolution to design efficient 1-Bit Convolutional Neural Networks (CNNs)
Our objective is to come up with a tiny yet efficient binary neural architecture by exploring the best candidates of the group convolution.
arXiv Detail & Related papers (2020-05-13T13:25:51Z) - Training Binary Neural Networks with Real-to-Binary Convolutions [52.91164959767517]
We show how to train binary networks to within a few percent points of the full precision counterpart.
We show how to build a strong baseline, which already achieves state-of-the-art accuracy.
We show that, when putting all of our improvements together, the proposed model beats the current state of the art by more than 5% top-1 accuracy on ImageNet.
arXiv Detail & Related papers (2020-03-25T17:54:38Z) - ReActNet: Towards Precise Binary Neural Network with Generalized
Activation Functions [76.05981545084738]
We propose several ideas for enhancing a binary network to close its accuracy gap from real-valued networks without incurring any additional computational cost.
We first construct a baseline network by modifying and binarizing a compact real-valued network with parameter-free shortcuts.
We show that the proposed ReActNet outperforms all the state-of-the-arts by a large margin.
arXiv Detail & Related papers (2020-03-07T02:12:02Z) - Exploring the Connection Between Binary and Spiking Neural Networks [1.329054857829016]
We bridge the recent algorithmic progress in training Binary Neural Networks and Spiking Neural Networks.
We show that training Spiking Neural Networks in the extreme quantization regime results in near full precision accuracies on large-scale datasets.
arXiv Detail & Related papers (2020-02-24T03:46:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.