IR2Net: Information Restriction and Information Recovery for Accurate
Binary Neural Networks
- URL: http://arxiv.org/abs/2210.02637v1
- Date: Thu, 6 Oct 2022 02:03:26 GMT
- Title: IR2Net: Information Restriction and Information Recovery for Accurate
Binary Neural Networks
- Authors: Ping Xue, Yang Lu, Jingfei Chang, Xing Wei, Zhen Wei
- Abstract summary: Weight and activation binarization can efficiently compress deep neural networks and accelerate model inference, but cause severe accuracy degradation.
We propose IR$2$Net to stimulate the potential of BNNs and improve the network accuracy by restricting the input information and recovering the feature information.
Experimental results demonstrate that our approach still achieves comparable accuracy even with $ sim $10x floating-point operations (FLOPs) reduction for ResNet-18.
- Score: 24.42067007684169
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Weight and activation binarization can efficiently compress deep neural
networks and accelerate model inference, but cause severe accuracy degradation.
Existing optimization methods for binary neural networks (BNNs) focus on
fitting full-precision networks to reduce quantization errors, and suffer from
the trade-off between accuracy and computational complexity. In contrast,
considering the limited learning ability and information loss caused by the
limited representational capability of BNNs, we propose IR$^2$Net to stimulate
the potential of BNNs and improve the network accuracy by restricting the input
information and recovering the feature information, including: 1) information
restriction: for a BNN, by evaluating the learning ability on the input
information, discarding some of the information it cannot focus on, and
limiting the amount of input information to match its learning ability; 2)
information recovery: due to the information loss in forward propagation, the
output feature information of the network is not enough to support accurate
classification. By selecting some shallow feature maps with richer information,
and fusing them with the final feature maps to recover the feature information.
In addition, the computational cost is reduced by streamlining the information
recovery method to strike a better trade-off between accuracy and efficiency.
Experimental results demonstrate that our approach still achieves comparable
accuracy even with $ \sim $10x floating-point operations (FLOPs) reduction for
ResNet-18. The models and code are available at
https://github.com/pingxue-hfut/IR2Net.
Related papers
- Rethinking Deep Learning: Propagating Information in Neural Networks without Backpropagation and Statistical Optimization [0.0]
This study discusses the information propagation capabilities and potential practical applications of NNs as neural system mimicking structures.
In this study, the NNs architecture comprises fully connected layers using step functions as activation functions, with 0-15 hidden layers, and no weight updates.
The accuracy is calculated by comparing the average output vectors of the training data for each label with the output vectors of the test data, based on vector similarity.
arXiv Detail & Related papers (2024-08-18T09:22:24Z) - BiHRNet: A Binary high-resolution network for Human Pose Estimation [11.250422970707415]
We propose a binary human pose estimator named BiHRNet, whose weights and activations are expressed as $pm$1.
BiHRNet retains the keypoint extraction ability of HRNet, while using fewer computing resources by adapting binary neural network (BNN)
We show BiHRNet achieves a PCKh of 87.9 on the MPII dataset, which outperforms all binary pose estimation networks.
arXiv Detail & Related papers (2023-11-17T03:01:37Z) - Network Binarization via Contrastive Learning [16.274341164897827]
We establish a novel contrastive learning framework while training Binary Neural Networks (BNNs)
MI is introduced as the metric to measure the information shared between binary and FP activations.
Results show that our method can be implemented as a pile-up module on existing state-of-the-art binarization methods.
arXiv Detail & Related papers (2022-07-06T21:04:53Z) - Distribution-sensitive Information Retention for Accurate Binary Neural
Network [49.971345958676196]
We present a novel Distribution-sensitive Information Retention Network (DIR-Net) to retain the information of the forward activations and backward gradients.
Our DIR-Net consistently outperforms the SOTA binarization approaches under mainstream and compact architectures.
We conduct our DIR-Net on real-world resource-limited devices which achieves 11.1 times storage saving and 5.4 times speedup.
arXiv Detail & Related papers (2021-09-25T10:59:39Z) - SignalNet: A Low Resolution Sinusoid Decomposition and Estimation
Network [79.04274563889548]
We propose SignalNet, a neural network architecture that detects the number of sinusoids and estimates their parameters from quantized in-phase and quadrature samples.
We introduce a worst-case learning threshold for comparing the results of our network relative to the underlying data distributions.
In simulation, we find that our algorithm is always able to surpass the threshold for three-bit data but often cannot exceed the threshold for one-bit data.
arXiv Detail & Related papers (2021-06-10T04:21:20Z) - S2-BNN: Bridging the Gap Between Self-Supervised Real and 1-bit Neural
Networks via Guided Distribution Calibration [74.5509794733707]
We present a novel guided learning paradigm from real-valued to distill binary networks on the final prediction distribution.
Our proposed method can boost the simple contrastive learning baseline by an absolute gain of 5.515% on BNNs.
Our method achieves substantial improvement over the simple contrastive learning baseline, and is even comparable to many mainstream supervised BNN methods.
arXiv Detail & Related papers (2021-02-17T18:59:28Z) - Binary Neural Networks: A Survey [126.67799882857656]
The binary neural network serves as a promising technique for deploying deep models on resource-limited devices.
The binarization inevitably causes severe information loss, and even worse, its discontinuity brings difficulty to the optimization of the deep network.
We present a survey of these algorithms, mainly categorized into the native solutions directly conducting binarization, and the optimized ones using techniques like minimizing the quantization error, improving the network loss function, and reducing the gradient error.
arXiv Detail & Related papers (2020-03-31T16:47:20Z) - BiDet: An Efficient Binarized Object Detector [96.19708396510894]
We propose a binarized neural network learning method called BiDet for efficient object detection.
Our BiDet fully utilizes the representational capacity of the binary neural networks for object detection by redundancy removal.
Our method outperforms the state-of-the-art binary neural networks by a sizable margin.
arXiv Detail & Related papers (2020-03-09T08:16:16Z) - Widening and Squeezing: Towards Accurate and Efficient QNNs [125.172220129257]
Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters.
Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques.
We address this problem by projecting features in original full-precision networks to high-dimensional quantization features.
arXiv Detail & Related papers (2020-02-03T04:11:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.