SoFAr: Shortcut-based Fractal Architectures for Binary Convolutional
Neural Networks
- URL: http://arxiv.org/abs/2009.05317v1
- Date: Fri, 11 Sep 2020 10:00:47 GMT
- Title: SoFAr: Shortcut-based Fractal Architectures for Binary Convolutional
Neural Networks
- Authors: Zhu Baozhou, Peter Hofstee, Jinho Lee, Zaid Al-Ars
- Abstract summary: We propose two Shortcut-based Fractal Architectures (SoFAr) specifically designed for Binary Convolutional Neural Networks (BCNNs)
Our proposed SoFAr combines the adoption of shortcuts and the fractal architectures in one unified model, which is helpful in the training of BCNNs.
Results show that our proposed SoFAr achieves better accuracy compared with shortcut-based BCNNs.
- Score: 7.753767947048147
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Binary Convolutional Neural Networks (BCNNs) can significantly improve the
efficiency of Deep Convolutional Neural Networks (DCNNs) for their deployment
on resource-constrained platforms, such as mobile and embedded systems.
However, the accuracy degradation of BCNNs is still considerable compared with
their full precision counterpart, impeding their practical deployment. Because
of the inevitable binarization error in the forward propagation and gradient
mismatch problem in the backward propagation, it is nontrivial to train BCNNs
to achieve satisfactory accuracy. To ease the difficulty of training, the
shortcut-based BCNNs, such as residual connection-based Bi-real ResNet and
dense connection-based BinaryDenseNet, introduce additional shortcuts in
addition to the shortcuts already present in their full precision counterparts.
Furthermore, fractal architectures have been also been used to improve the
training process of full-precision DCNNs since the fractal structure triggers
effects akin to deep supervision and lateral student-teacher information flow.
Inspired by the shortcuts and fractal architectures, we propose two
Shortcut-based Fractal Architectures (SoFAr) specifically designed for BCNNs:
1. residual connection-based fractal architectures for binary ResNet, and 2.
dense connection-based fractal architectures for binary DenseNet. Our proposed
SoFAr combines the adoption of shortcuts and the fractal architectures in one
unified model, which is helpful in the training of BCNNs. Results show that our
proposed SoFAr achieves better accuracy compared with shortcut-based BCNNs.
Specifically, the Top-1 accuracy of our proposed RF-c4d8 ResNet37(41) and
DRF-c2d2 DenseNet51(53) on ImageNet outperforms Bi-real ResNet18(64) and
BinaryDenseNet51(32) by 3.29% and 1.41%, respectively, with the same
computational complexity overhead.
Related papers
- Binary domain generalization for sparsifying binary neural networks [3.2462411268263964]
Binary neural networks (BNNs) are an attractive solution for developing and deploying deep neural network (DNN)-based applications in resource constrained devices.
Weight pruning of BNNs leads to performance degradation, which suggests that the standard binarization domain of BNNs is not well adapted for the task.
This work proposes a novel more general binary domain that extends the standard binary one that is more robust to pruning techniques.
arXiv Detail & Related papers (2023-06-23T14:32:16Z) - BiFSMNv2: Pushing Binary Neural Networks for Keyword Spotting to
Real-Network Performance [54.214426436283134]
Deep neural networks, such as the Deep-FSMN, have been widely studied for keyword spotting (KWS) applications.
We present a strong yet efficient binary neural network for KWS, namely BiFSMNv2, pushing it to the real-network accuracy performance.
We highlight that benefiting from the compact architecture and optimized hardware kernel, BiFSMNv2 can achieve an impressive 25.1x speedup and 20.2x storage-saving on edge hardware.
arXiv Detail & Related papers (2022-11-13T18:31:45Z) - Basic Binary Convolution Unit for Binarized Image Restoration Network [146.0988597062618]
In this study, we reconsider components in binary convolution, such as residual connection, BatchNorm, activation function, and structure, for image restoration tasks.
Based on our findings and analyses, we design a simple yet efficient basic binary convolution unit (BBCU)
Our BBCU significantly outperforms other BNNs and lightweight models, which shows that BBCU can serve as a basic unit for binarized IR networks.
arXiv Detail & Related papers (2022-10-02T01:54:40Z) - BiFSMN: Binary Neural Network for Keyword Spotting [47.46397208920726]
BiFSMN is an accurate and extreme-efficient binary neural network for KWS.
We show that BiFSMN can achieve an impressive 22.3x speedup and 15.5x storage-saving on real-world edge hardware.
arXiv Detail & Related papers (2022-02-14T05:16:53Z) - Elastic-Link for Binarized Neural Network [9.83865304744923]
"Elastic-Link" (EL) module enrich information flow within a BNN by adaptively adding real-valued input features to the subsequent convolutional output features.
EL produces a significant improvement on the challenging large-scale ImageNet dataset.
With the integration of ReActNet, it yields a new state-of-the-art result of 71.9% top-1 accuracy.
arXiv Detail & Related papers (2021-12-19T13:49:29Z) - Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
Neural Networks [72.81092567651395]
Sub-bit Neural Networks (SNNs) are a new type of binary quantization design tailored to compress and accelerate BNNs.
SNNs are trained with a kernel-aware optimization framework, which exploits binary quantization in the fine-grained convolutional kernel space.
Experiments on visual recognition benchmarks and the hardware deployment on FPGA validate the great potentials of SNNs.
arXiv Detail & Related papers (2021-10-18T11:30:29Z) - Structured Convolutions for Efficient Neural Network Design [65.36569572213027]
We tackle model efficiency by exploiting redundancy in the textitimplicit structure of the building blocks of convolutional neural networks.
We show how this decomposition can be applied to 2D and 3D kernels as well as the fully-connected layers.
arXiv Detail & Related papers (2020-08-06T04:38:38Z) - Distillation Guided Residual Learning for Binary Convolutional Neural
Networks [83.6169936912264]
It is challenging to bridge the performance gap between Binary CNN (BCNN) and Floating point CNN (FCNN)
We observe that, this performance gap leads to substantial residuals between intermediate feature maps of BCNN and FCNN.
To minimize the performance gap, we enforce BCNN to produce similar intermediate feature maps with the ones of FCNN.
This training strategy, i.e., optimizing each binary convolutional block with block-wise distillation loss derived from FCNN, leads to a more effective optimization to BCNN.
arXiv Detail & Related papers (2020-07-10T07:55:39Z) - Binarizing MobileNet via Evolution-based Searching [66.94247681870125]
We propose a use of evolutionary search to facilitate the construction and training scheme when binarizing MobileNet.
Inspired by one-shot architecture search frameworks, we manipulate the idea of group convolution to design efficient 1-Bit Convolutional Neural Networks (CNNs)
Our objective is to come up with a tiny yet efficient binary neural architecture by exploring the best candidates of the group convolution.
arXiv Detail & Related papers (2020-05-13T13:25:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.