An Optical XNOR-Bitcount Based Accelerator for Efficient Inference of
Binary Neural Networks
- URL: http://arxiv.org/abs/2302.06405v2
- Date: Mon, 20 Mar 2023 02:34:18 GMT
- Title: An Optical XNOR-Bitcount Based Accelerator for Efficient Inference of
Binary Neural Networks
- Authors: Sairam Sri Vatsavai, Venkata Sai Praneeth Karempudi, and Ishan Thakkar
- Abstract summary: We invent a single-MRR-based optical XNOR gate (OXG)
We present a novel design of bitcount circuit which we refer to as Photo-Charge Accumulator (PCA)
Our evaluation for the inference of four modern BNNs indicates that OXBNN provides improvements of up to 62x and 7.6x in frames-per-second (FPS) and FPS/W (energy efficiency)
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Binary Neural Networks (BNNs) are increasingly preferred over full-precision
Convolutional Neural Networks(CNNs) to reduce the memory and computational
requirements of inference processing with minimal accuracy drop. BNNs convert
CNN model parameters to 1-bit precision, allowing inference of BNNs to be
processed with simple XNOR and bitcount operations. This makes BNNs amenable to
hardware acceleration. Several photonic integrated circuits (PICs) based BNN
accelerators have been proposed. Although these accelerators provide remarkably
higher throughput and energy efficiency than their electronic counterparts, the
utilized XNOR and bitcount circuits in these accelerators need to be further
enhanced to improve their area, energy efficiency, and throughput. This paper
aims to fulfill this need. For that, we invent a single-MRR-based optical XNOR
gate (OXG). Moreover, we present a novel design of bitcount circuit which we
refer to as Photo-Charge Accumulator (PCA). We employ multiple OXGs in a
cascaded manner using dense wavelength division multiplexing (DWDM) and connect
them to the PCA, to forge a novel Optical XNOR-Bitcount based Binary Neural
Network Accelerator (OXBNN). Our evaluation for the inference of four modern
BNNs indicates that OXBNN provides improvements of up to 62x and 7.6x in
frames-per-second (FPS) and FPS/W (energy efficiency), respectively, on
geometric mean over two PIC-based BNN accelerators from prior work. We
developed a transaction-level, event-driven python-based simulator for
evaluation of accelerators (https://github.com/uky-UCAT/B_ONN_SIM).
Related papers
- Compacting Binary Neural Networks by Sparse Kernel Selection [58.84313343190488]
This paper is motivated by a previously revealed phenomenon that the binary kernels in successful BNNs are nearly power-law distributed.
We develop the Permutation Straight-Through Estimator (PSTE) that is able to not only optimize the selection process end-to-end but also maintain the non-repetitive occupancy of selected codewords.
Experiments verify that our method reduces both the model size and bit-wise computational costs, and achieves accuracy improvements compared with state-of-the-art BNNs under comparable budgets.
arXiv Detail & Related papers (2023-03-25T13:53:02Z) - Basic Binary Convolution Unit for Binarized Image Restoration Network [146.0988597062618]
In this study, we reconsider components in binary convolution, such as residual connection, BatchNorm, activation function, and structure, for image restoration tasks.
Based on our findings and analyses, we design a simple yet efficient basic binary convolution unit (BBCU)
Our BBCU significantly outperforms other BNNs and lightweight models, which shows that BBCU can serve as a basic unit for binarized IR networks.
arXiv Detail & Related papers (2022-10-02T01:54:40Z) - Recurrent Bilinear Optimization for Binary Neural Networks [58.972212365275595]
BNNs neglect the intrinsic bilinear relationship of real-valued weights and scale factors.
Our work is the first attempt to optimize BNNs from the bilinear perspective.
We obtain robust RBONNs, which show impressive performance over state-of-the-art BNNs on various models and datasets.
arXiv Detail & Related papers (2022-09-04T06:45:33Z) - Binary Neural Networks as a general-propose compute paradigm for
on-device computer vision [0.0]
We propose a BNN framework comprising 1) a minimalistic inference scheme for hardware-friendliness, 2) an over- parameterized training scheme for high accuracy, and 3) a simple procedure to adapt to different vision tasks.
The resultant framework overtakes 8-bit quantization in the speed-vs-accuracy tradeoff for classification, detection, segmentation, super-resolution and matching.
Our BNNs promise 2.8-7$times$ fewer execution cycles than 8-bit and 2.1-2.7$times$ fewer cycles than alternative BNN designs.
arXiv Detail & Related papers (2022-02-08T08:38:22Z) - Sub-bit Neural Networks: Learning to Compress and Accelerate Binary
Neural Networks [72.81092567651395]
Sub-bit Neural Networks (SNNs) are a new type of binary quantization design tailored to compress and accelerate BNNs.
SNNs are trained with a kernel-aware optimization framework, which exploits binary quantization in the fine-grained convolutional kernel space.
Experiments on visual recognition benchmarks and the hardware deployment on FPGA validate the great potentials of SNNs.
arXiv Detail & Related papers (2021-10-18T11:30:29Z) - ROBIN: A Robust Optical Binary Neural Network Accelerator [3.8137985834223507]
Domain specific neural network accelerators have garnered attention because of their improved energy efficiency and inference performance.
We present a novel optical-domain BNN accelerator, named ROBIN, which intelligently integrates heterogeneous microring resonator optical devices.
Our analysis shows that ROBIN can outperform the best-known optical BNN accelerators and also many electronic accelerators.
arXiv Detail & Related papers (2021-07-12T16:00:32Z) - Quantized Neural Networks via {-1, +1} Encoding Decomposition and
Acceleration [83.84684675841167]
We propose a novel encoding scheme using -1, +1 to decompose quantized neural networks (QNNs) into multi-branch binary networks.
We validate the effectiveness of our method on large-scale image classification, object detection, and semantic segmentation tasks.
arXiv Detail & Related papers (2021-06-18T03:11:15Z) - FracBNN: Accurate and FPGA-Efficient Binary Neural Networks with
Fractional Activations [20.218382369944152]
Binary neural networks (BNNs) have 1-bit weights and activations.
BNNs tend to produce a much lower accuracy on realistic datasets such as ImageNet.
This work proposes FracBNN, which exploits fractional activations to substantially improve the accuracy of BNNs.
arXiv Detail & Related papers (2020-12-22T17:49:30Z) - FTBNN: Rethinking Non-linearity for 1-bit CNNs and Going Beyond [23.5996182207431]
We show that binarized convolution process owns an increasing linearity towards the target of minimizing such error, which in turn hampers BNN's discriminative ability.
We re-investigate and tune proper non-linear modules to fix that contradiction, leading to a strong baseline which achieves state-of-the-art performance.
arXiv Detail & Related papers (2020-10-19T08:11:48Z) - Binarized Graph Neural Network [65.20589262811677]
We develop a binarized graph neural network to learn the binary representations of the nodes with binary network parameters.
Our proposed method can be seamlessly integrated into the existing GNN-based embedding approaches.
Experiments indicate that the proposed binarized graph neural network, namely BGN, is orders of magnitude more efficient in terms of both time and space.
arXiv Detail & Related papers (2020-04-19T09:43:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.