NASB: Neural Architecture Search for Binary Convolutional Neural
Networks
- URL: http://arxiv.org/abs/2008.03515v1
- Date: Sat, 8 Aug 2020 13:06:11 GMT
- Title: NASB: Neural Architecture Search for Binary Convolutional Neural
Networks
- Authors: Baozhou Zhu, Zaid Al-Ars, Peter Hofstee
- Abstract summary: We propose a strategy, named NASB, which adopts Neural Architecture Search (NAS) to find an optimal architecture for the binarization of CNNs.
Due to the flexibility of this automated strategy, the obtained architecture is not only suitable for binarization but also has low overhead.
NASB outperforms existing single and multiple binary CNNs by up to 4.0% and 1.0% Top-1 accuracy respectively.
- Score: 2.3204178451683264
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Binary Convolutional Neural Networks (CNNs) have significantly reduced the
number of arithmetic operations and the size of memory storage needed for CNNs,
which makes their deployment on mobile and embedded systems more feasible.
However, the CNN architecture after binarizing requires to be redesigned and
refined significantly due to two reasons: 1. the large accumulation error of
binarization in the forward propagation, and 2. the severe gradient mismatch
problem of binarization in the backward propagation. Even though the
substantial effort has been invested in designing architectures for single and
multiple binary CNNs, it is still difficult to find an optimal architecture for
binary CNNs. In this paper, we propose a strategy, named NASB, which adopts
Neural Architecture Search (NAS) to find an optimal architecture for the
binarization of CNNs. Due to the flexibility of this automated strategy, the
obtained architecture is not only suitable for binarization but also has low
overhead, achieving a better trade-off between the accuracy and computational
complexity of hand-optimized binary CNNs. The implementation of NASB strategy
is evaluated on the ImageNet dataset and demonstrated as a better solution
compared to existing quantized CNNs. With the insignificant overhead increase,
NASB outperforms existing single and multiple binary CNNs by up to 4.0% and
1.0% Top-1 accuracy respectively, bringing them closer to the precision of
their full precision counterpart. The code and pretrained models will be
publicly available.
Related papers
- BinaryViT: Pushing Binary Vision Transformers Towards Convolutional
Models [0.0]
Binarization can be used to help reduce the size of ViT models and their computational cost significantly.
ViTs suffer a larger performance drop when directly applying convolutional neural network (CNN) binarization methods.
We propose BinaryViT, in which inspired by the CNN architecture, we include operations from the CNN architecture into a pure ViT architecture.
arXiv Detail & Related papers (2023-06-29T04:48:02Z) - Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
Neural Networks [72.81092567651395]
Sub-bit Neural Networks (SNNs) are a new type of binary quantization design tailored to compress and accelerate BNNs.
SNNs are trained with a kernel-aware optimization framework, which exploits binary quantization in the fine-grained convolutional kernel space.
Experiments on visual recognition benchmarks and the hardware deployment on FPGA validate the great potentials of SNNs.
arXiv Detail & Related papers (2021-10-18T11:30:29Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - BARS: Joint Search of Cell Topology and Layout for Accurate and
Efficient Binary ARchitectures [21.671696519808226]
Binary Neural Networks (BNNs) have received significant attention due to their promising efficiency.
Currently, most BNN studies directly adopt widely-used CNN architectures, which can be suboptimal for BNNs.
This paper proposes a novel Binary ARchitecture Search (BARS) flow to discover superior binary architecture in a large design space.
arXiv Detail & Related papers (2020-11-21T14:38:44Z) - High-Capacity Expert Binary Networks [56.87581500474093]
Network binarization is a promising hardware-aware direction for creating efficient deep models.
Despite its memory and computational advantages, reducing the accuracy gap between binary models and their real-valued counterparts remains an unsolved challenging research problem.
We propose Expert Binary Convolution, which, for the first time, tailors conditional computing to binary networks by learning to select one data-specific expert binary filter at a time conditioned on input features.
arXiv Detail & Related papers (2020-10-07T17:58:10Z) - FATNN: Fast and Accurate Ternary Neural Networks [89.07796377047619]
Ternary Neural Networks (TNNs) have received much attention due to being potentially orders of magnitude faster in inference, as well as more power efficient, than full-precision counterparts.
In this work, we show that, under some mild constraints, computational complexity of the ternary inner product can be reduced by a factor of 2.
We elaborately design an implementation-dependent ternary quantization algorithm to mitigate the performance gap.
arXiv Detail & Related papers (2020-08-12T04:26:18Z) - Towards Lossless Binary Convolutional Neural Networks Using Piecewise
Approximation [4.023728681102073]
CNNs can significantly reduce the number of arithmetic operations and the size of memory storage.
However, the accuracy degradation of single and multiple binary CNNs is unacceptable for modern architectures.
We propose a Piecewise Approximation scheme for multiple binary CNNs which lessens accuracy loss by approximating full precision weights and activations.
arXiv Detail & Related papers (2020-08-08T13:32:33Z) - Binarizing MobileNet via Evolution-based Searching [66.94247681870125]
We propose a use of evolutionary search to facilitate the construction and training scheme when binarizing MobileNet.
Inspired by one-shot architecture search frameworks, we manipulate the idea of group convolution to design efficient 1-Bit Convolutional Neural Networks (CNNs)
Our objective is to come up with a tiny yet efficient binary neural architecture by exploring the best candidates of the group convolution.
arXiv Detail & Related papers (2020-05-13T13:25:51Z) - Widening and Squeezing: Towards Accurate and Efficient QNNs [125.172220129257]
Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters.
Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques.
We address this problem by projecting features in original full-precision networks to high-dimensional quantization features.
arXiv Detail & Related papers (2020-02-03T04:11:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.