Binarizing MobileNet via Evolution-based Searching
- URL: http://arxiv.org/abs/2005.06305v2
- Date: Fri, 15 May 2020 15:48:58 GMT
- Title: Binarizing MobileNet via Evolution-based Searching
- Authors: Hai Phan, Zechun Liu, Dang Huynh, Marios Savvides, Kwang-Ting Cheng,
Zhiqiang Shen
- Abstract summary: We propose a use of evolutionary search to facilitate the construction and training scheme when binarizing MobileNet.
Inspired by one-shot architecture search frameworks, we manipulate the idea of group convolution to design efficient 1-Bit Convolutional Neural Networks (CNNs)
Our objective is to come up with a tiny yet efficient binary neural architecture by exploring the best candidates of the group convolution.
- Score: 66.94247681870125
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Binary Neural Networks (BNNs), known to be one among the effectively compact
network architectures, have achieved great outcomes in the visual tasks.
Designing efficient binary architectures is not trivial due to the binary
nature of the network. In this paper, we propose a use of evolutionary search
to facilitate the construction and training scheme when binarizing MobileNet, a
compact network with separable depth-wise convolution. Inspired by one-shot
architecture search frameworks, we manipulate the idea of group convolution to
design efficient 1-Bit Convolutional Neural Networks (CNNs), assuming an
approximately optimal trade-off between computational cost and model accuracy.
Our objective is to come up with a tiny yet efficient binary neural
architecture by exploring the best candidates of the group convolution while
optimizing the model performance in terms of complexity and latency. The
approach is threefold. First, we train strong baseline binary networks with a
wide range of random group combinations at each convolutional layer. This
set-up gives the binary neural networks a capability of preserving essential
information through layers. Second, to find a good set of hyperparameters for
group convolutions we make use of the evolutionary search which leverages the
exploration of efficient 1-bit models. Lastly, these binary models are trained
from scratch in a usual manner to achieve the final binary model. Various
experiments on ImageNet are conducted to show that following our construction
guideline, the final model achieves 60.09% Top-1 accuracy and outperforms the
state-of-the-art CI-BCNN with the same computational cost.
Related papers
- Tiled Bit Networks: Sub-Bit Neural Network Compression Through Reuse of Learnable Binary Vectors [4.95475852994362]
We propose a new form of quantization to tile neural network layers with sequences of bits to achieve sub-bit compression of binary-weighted neural networks.
We employ the approach to both fully-connected and convolutional layers, which make up the breadth of space in most neural architectures.
arXiv Detail & Related papers (2024-07-16T15:55:38Z) - Compacting Binary Neural Networks by Sparse Kernel Selection [58.84313343190488]
This paper is motivated by a previously revealed phenomenon that the binary kernels in successful BNNs are nearly power-law distributed.
We develop the Permutation Straight-Through Estimator (PSTE) that is able to not only optimize the selection process end-to-end but also maintain the non-repetitive occupancy of selected codewords.
Experiments verify that our method reduces both the model size and bit-wise computational costs, and achieves accuracy improvements compared with state-of-the-art BNNs under comparable budgets.
arXiv Detail & Related papers (2023-03-25T13:53:02Z) - Join the High Accuracy Club on ImageNet with A Binary Neural Network
Ticket [10.552465253379134]
We focus on a problem: how can a binary neural network achieve the crucial accuracy level (e.g., 80%) on ILSVRC-2012 ImageNet?
We design a novel binary architecture BNext based on a comprehensive study of binary architectures and their optimization process.
We propose a novel knowledge-distillation technique to alleviate the counter-intuitive overfitting problem observed when attempting to train extremely accurate binary models.
arXiv Detail & Related papers (2022-11-23T13:08:58Z) - Towards Accurate Binary Neural Networks via Modeling Contextual
Dependencies [52.691032025163175]
Existing Binary Neural Networks (BNNs) operate mainly on local convolutions with binarization function.
We present new designs of binary neural modules, which enables leading binary neural modules by a large margin.
arXiv Detail & Related papers (2022-09-03T11:51:04Z) - Differentiable Neural Architecture Learning for Efficient Neural Network
Design [31.23038136038325]
We introduce a novel emph architecture parameterisation based on scaled sigmoid function.
We then propose a general emphiable Neural Architecture Learning (DNAL) method to optimize the neural architecture without the need to evaluate candidate neural networks.
arXiv Detail & Related papers (2021-03-03T02:03:08Z) - Binary Graph Neural Networks [69.51765073772226]
Graph Neural Networks (GNNs) have emerged as a powerful and flexible framework for representation learning on irregular data.
In this paper, we present and evaluate different strategies for the binarization of graph neural networks.
We show that through careful design of the models, and control of the training process, binary graph neural networks can be trained at only a moderate cost in accuracy on challenging benchmarks.
arXiv Detail & Related papers (2020-12-31T18:48:58Z) - SAR-NAS: Skeleton-based Action Recognition via Neural Architecture
Searching [18.860051578038608]
We encode a skeleton-based action instance into a tensor and define a set of operations to build two types of network cells: normal cells and reduction cells.
Experiments on the challenging NTU RGB+D and Kinectics datasets have verified that most of the networks developed to date for skeleton-based action recognition are likely not compact and efficient.
The proposed method provides an approach to search for such a compact network that is able to achieve comparative or even better performance than the state-of-the-art methods.
arXiv Detail & Related papers (2020-10-29T03:24:15Z) - High-Capacity Expert Binary Networks [56.87581500474093]
Network binarization is a promising hardware-aware direction for creating efficient deep models.
Despite its memory and computational advantages, reducing the accuracy gap between binary models and their real-valued counterparts remains an unsolved challenging research problem.
We propose Expert Binary Convolution, which, for the first time, tailors conditional computing to binary networks by learning to select one data-specific expert binary filter at a time conditioned on input features.
arXiv Detail & Related papers (2020-10-07T17:58:10Z) - NASB: Neural Architecture Search for Binary Convolutional Neural
Networks [2.3204178451683264]
We propose a strategy, named NASB, which adopts Neural Architecture Search (NAS) to find an optimal architecture for the binarization of CNNs.
Due to the flexibility of this automated strategy, the obtained architecture is not only suitable for binarization but also has low overhead.
NASB outperforms existing single and multiple binary CNNs by up to 4.0% and 1.0% Top-1 accuracy respectively.
arXiv Detail & Related papers (2020-08-08T13:06:11Z) - Structured Convolutions for Efficient Neural Network Design [65.36569572213027]
We tackle model efficiency by exploiting redundancy in the textitimplicit structure of the building blocks of convolutional neural networks.
We show how this decomposition can be applied to 2D and 3D kernels as well as the fully-connected layers.
arXiv Detail & Related papers (2020-08-06T04:38:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.