SONIC: A Sparse Neural Network Inference Accelerator with Silicon
Photonics for Energy-Efficient Deep Learning
- URL: http://arxiv.org/abs/2109.04459v1
- Date: Thu, 9 Sep 2021 17:57:09 GMT
- Title: SONIC: A Sparse Neural Network Inference Accelerator with Silicon
Photonics for Energy-Efficient Deep Learning
- Authors: Febin Sunny, Mahdi Nikdast, Sudeep Pasricha
- Abstract summary: We propose a novel silicon photonics-based sparse neural network inference accelerator called SONIC.
SONIC can achieve up to 5.8x better performance-per-watt and 8.4x lower energy-per-bit than state-of-the-art sparse electronic neural network accelerators.
- Score: 4.286327408435937
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Sparse neural networks can greatly facilitate the deployment of neural
networks on resource-constrained platforms as they offer compact model sizes
while retaining inference accuracy. Because of the sparsity in parameter
matrices, sparse neural networks can, in principle, be exploited in accelerator
architectures for improved energy-efficiency and latency. However, to realize
these improvements in practice, there is a need to explore sparsity-aware
hardware-software co-design. In this paper, we propose a novel silicon
photonics-based sparse neural network inference accelerator called SONIC. Our
experimental analysis shows that SONIC can achieve up to 5.8x better
performance-per-watt and 8.4x lower energy-per-bit than state-of-the-art sparse
electronic neural network accelerators; and up to 13.8x better
performance-per-watt and 27.6x lower energy-per-bit than the best known
photonic neural network accelerators.
Related papers
- Single Neuromorphic Memristor closely Emulates Multiple Synaptic
Mechanisms for Energy Efficient Neural Networks [71.79257685917058]
We demonstrate memristive nano-devices based on SrTiO3 that inherently emulate all these synaptic functions.
These memristors operate in a non-filamentary, low conductance regime, which enables stable and energy efficient operation.
arXiv Detail & Related papers (2024-02-26T15:01:54Z) - SpikingJelly: An open-source machine learning infrastructure platform
for spike-based intelligence [51.6943465041708]
Spiking neural networks (SNNs) aim to realize brain-inspired intelligence on neuromorphic chips with high energy efficiency.
We contribute a full-stack toolkit for pre-processing neuromorphic datasets, building deep SNNs, optimizing their parameters, and deploying SNNs on neuromorphic chips.
arXiv Detail & Related papers (2023-10-25T13:15:17Z) - Spatially Varying Nanophotonic Neural Networks [39.1303097259564]
Photonic processors that execute operations using photons instead of electrons promise to enable optical neural networks with ultra-low latency and power consumption.
Existing optical neural networks, limited by the underlying network designs, have achieved image recognition accuracy far below that of state-of-the-art electronic neural networks.
arXiv Detail & Related papers (2023-08-07T08:48:46Z) - A Hybrid Neural Coding Approach for Pattern Recognition with Spiking
Neural Networks [53.31941519245432]
Brain-inspired spiking neural networks (SNNs) have demonstrated promising capabilities in solving pattern recognition tasks.
These SNNs are grounded on homogeneous neurons that utilize a uniform neural coding for information representation.
In this study, we argue that SNN architectures should be holistically designed to incorporate heterogeneous coding schemes.
arXiv Detail & Related papers (2023-05-26T02:52:12Z) - A Resource-efficient Spiking Neural Network Accelerator Supporting
Emerging Neural Encoding [6.047137174639418]
Spiking neural networks (SNNs) recently gained momentum due to their low-power multiplication-free computing.
SNNs require very long spike trains (up to 1000) to reach an accuracy similar to their artificial neural network (ANN) counterparts for large models.
We present a novel hardware architecture that can efficiently support SNN with emerging neural encoding.
arXiv Detail & Related papers (2022-06-06T10:56:25Z) - Optimizing the Consumption of Spiking Neural Networks with Activity
Regularization [15.317534913990633]
Spiking Neural Networks (SNNs) are an example of bio-inspired techniques that can further save energy by using binary activations, and avoid consuming energy when not spiking.
In this work, we look into different techniques to enforce sparsity on the neural network activation maps and compare the effect of different training regularizers on the efficiency of the optimized DNNs and SNNs.
arXiv Detail & Related papers (2022-04-04T13:19:47Z) - FPGA-optimized Hardware acceleration for Spiking Neural Networks [69.49429223251178]
This work presents the development of a hardware accelerator for an SNN, with off-line training, applied to an image recognition task.
The design targets a Xilinx Artix-7 FPGA, using in total around the 40% of the available hardware resources.
It reduces the classification time by three orders of magnitude, with a small 4.5% impact on the accuracy, if compared to its software, full precision counterpart.
arXiv Detail & Related papers (2022-01-18T13:59:22Z) - Two Sparsities Are Better Than One: Unlocking the Performance Benefits
of Sparse-Sparse Networks [0.0]
We introduce Complementary Sparsity, a technique that significantly improves the performance of dual sparse networks on existing hardware.
We show up to 100X improvement in throughput and energy efficiency performing inference on FPGAs.
Our results suggest that weight plus activation sparsity can be a potent combination for efficiently scaling future AI models.
arXiv Detail & Related papers (2021-12-27T20:41:01Z) - Silicon photonic subspace neural chip for hardware-efficient deep
learning [11.374005508708995]
optical neural network (ONN) is a promising candidate for next-generation neurocomputing.
We devise a hardware-efficient photonic subspace neural network architecture.
We experimentally demonstrate our PSNN on a butterfly-style programmable silicon photonic integrated circuit.
arXiv Detail & Related papers (2021-11-11T06:34:05Z) - Building Compact and Robust Deep Neural Networks with Toeplitz Matrices [93.05076144491146]
This thesis focuses on the problem of training neural networks which are compact, easy to train, reliable and robust to adversarial examples.
We leverage the properties of structured matrices from the Toeplitz family to build compact and secure neural networks.
arXiv Detail & Related papers (2021-09-02T13:58:12Z) - AdderNet and its Minimalist Hardware Design for Energy-Efficient
Artificial Intelligence [111.09105910265154]
We present a novel minimalist hardware architecture using adder convolutional neural network (AdderNet)
The whole AdderNet can practically achieve 16% enhancement in speed.
We conclude the AdderNet is able to surpass all the other competitors.
arXiv Detail & Related papers (2021-01-25T11:31:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.