You Only Spike Once: Improving Energy-Efficient Neuromorphic Inference
to ANN-Level Accuracy
- URL: http://arxiv.org/abs/2006.09982v2
- Date: Sun, 8 Nov 2020 09:10:57 GMT
- Title: You Only Spike Once: Improving Energy-Efficient Neuromorphic Inference
to ANN-Level Accuracy
- Authors: Srivatsa P and Kyle Timothy Ng Chu and Burin Amornpaisannon and
Yaswanth Tavva and Venkata Pavan Kumar Miriyala and Jibin Wu and Malu Zhang
and Haizhou Li and Trevor E. Carlson
- Abstract summary: Spiking Neural Networks (SNNs) are a type of neuromorphic, or brain-inspired network.
SNNs are sparse, accessing very few weights, and typically only use addition operations instead of the more power-intensive multiply-and-accumulate operations.
In this work, we aim to overcome the limitations of TTFS-encoded neuromorphic systems.
- Score: 51.861168222799186
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the past decade, advances in Artificial Neural Networks (ANNs) have
allowed them to perform extremely well for a wide range of tasks. In fact, they
have reached human parity when performing image recognition, for example.
Unfortunately, the accuracy of these ANNs comes at the expense of a large
number of cache and/or memory accesses and compute operations. Spiking Neural
Networks (SNNs), a type of neuromorphic, or brain-inspired network, have
recently gained significant interest as power-efficient alternatives to ANNs,
because they are sparse, accessing very few weights, and typically only use
addition operations instead of the more power-intensive multiply-and-accumulate
(MAC) operations. The vast majority of neuromorphic hardware designs support
rate-encoded SNNs, where the information is encoded in spike rates.
Rate-encoded SNNs could be seen as inefficient as an encoding scheme because it
involves the transmission of a large number of spikes. A more efficient
encoding scheme, Time-To-First-Spike (TTFS) encoding, encodes information in
the relative time of arrival of spikes. While TTFS-encoded SNNs are more
efficient than rate-encoded SNNs, they have, up to now, performed poorly in
terms of accuracy compared to previous methods. Hence, in this work, we aim to
overcome the limitations of TTFS-encoded neuromorphic systems. To accomplish
this, we propose: (1) a novel optimization algorithm for TTFS-encoded SNNs
converted from ANNs and (2) a novel hardware accelerator for TTFS-encoded SNNs,
with a scalable and low-power design. Overall, our work in TTFS encoding and
training improves the accuracy of SNNs to achieve state-of-the-art results on
MNIST MLPs, while reducing power consumption by 1.46$\times$ over the
state-of-the-art neuromorphic hardware.
Related papers
- LC-TTFS: Towards Lossless Network Conversion for Spiking Neural Networks
with TTFS Coding [55.64533786293656]
We show that our algorithm can achieve a near-perfect mapping between the activation values of an ANN and the spike times of an SNN on a number of challenging AI tasks.
The study paves the way for deploying ultra-low-power TTFS-based SNNs on power-constrained edge computing platforms.
arXiv Detail & Related papers (2023-10-23T14:26:16Z) - Ultra-low-power Image Classification on Neuromorphic Hardware [3.784976081087904]
Spiking neural networks (SNNs) promise ultra-low-power applications by exploiting temporal and spatial sparsity.
Training such SNNs using backpropagation through time for vision tasks that rely mainly on spatial features is computationally costly.
We propose a temporal ANN-to-SNN conversion method, which we call Quartz, that is based on the time to first spike.
arXiv Detail & Related papers (2023-09-28T18:53:43Z) - A Hybrid Neural Coding Approach for Pattern Recognition with Spiking
Neural Networks [53.31941519245432]
Brain-inspired spiking neural networks (SNNs) have demonstrated promising capabilities in solving pattern recognition tasks.
These SNNs are grounded on homogeneous neurons that utilize a uniform neural coding for information representation.
In this study, we argue that SNN architectures should be holistically designed to incorporate heterogeneous coding schemes.
arXiv Detail & Related papers (2023-05-26T02:52:12Z) - Timing-Based Backpropagation in Spiking Neural Networks Without
Single-Spike Restrictions [2.8360662552057323]
We propose a novel backpropagation algorithm for training spiking neural networks (SNNs)
It encodes information in the relative multiple spike timing of individual neurons without single-spike restrictions.
arXiv Detail & Related papers (2022-11-29T11:38:33Z) - Training High-Performance Low-Latency Spiking Neural Networks by
Differentiation on Spike Representation [70.75043144299168]
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware.
It is a challenge to efficiently train SNNs due to their non-differentiability.
We propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance.
arXiv Detail & Related papers (2022-05-01T12:44:49Z) - Training Energy-Efficient Deep Spiking Neural Networks with Single-Spike
Hybrid Input Encoding [5.725845886457027]
Spiking Neural Networks (SNNs) provide higher computational efficiency in event driven neuromorphic hardware.
SNNs suffer from high inference latency, resulting from inefficient input encoding and training techniques.
This paper presents a training framework for low-latency energy-efficient SNNs.
arXiv Detail & Related papers (2021-07-26T06:16:40Z) - Sparse Spiking Gradient Descent [2.741266294612776]
We present the first sparse SNN backpropagation algorithm which achieves the same or better accuracy as current state of the art methods.
We show the effectiveness of our method on real datasets of varying complexity.
arXiv Detail & Related papers (2021-05-18T20:00:55Z) - FATNN: Fast and Accurate Ternary Neural Networks [89.07796377047619]
Ternary Neural Networks (TNNs) have received much attention due to being potentially orders of magnitude faster in inference, as well as more power efficient, than full-precision counterparts.
In this work, we show that, under some mild constraints, computational complexity of the ternary inner product can be reduced by a factor of 2.
We elaborately design an implementation-dependent ternary quantization algorithm to mitigate the performance gap.
arXiv Detail & Related papers (2020-08-12T04:26:18Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.