Towards Energy-Efficient, Low-Latency and Accurate Spiking LSTMs
- URL: http://arxiv.org/abs/2210.12613v1
- Date: Sun, 23 Oct 2022 04:10:27 GMT
- Title: Towards Energy-Efficient, Low-Latency and Accurate Spiking LSTMs
- Authors: Gourav Datta and Haoqin Deng and Robert Aviles and Peter A. Beerel
- Abstract summary: Spiking Neural Networks (SNNs) have emerged as an attractive-temporal computing paradigm vision for complex tasks.
We propose an optimized spiking long short-term memory networks (LSTM) training framework that involves a novel.
rev-to-SNN conversion framework, followed by SNN training.
We evaluate our framework on sequential learning tasks including temporal M, Google Speech Commands (GSC) datasets, and UCI Smartphone on different LSTM architectures.
- Score: 1.7969777786551424
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Spiking Neural Networks (SNNs) have emerged as an attractive spatio-temporal
computing paradigm for complex vision tasks. However, most existing works yield
models that require many time steps and do not leverage the inherent temporal
dynamics of spiking neural networks, even for sequential tasks. Motivated by
this observation, we propose an \rev{optimized spiking long short-term memory
networks (LSTM) training framework that involves a novel ANN-to-SNN conversion
framework, followed by SNN training}. In particular, we propose novel
activation functions in the source LSTM architecture and judiciously select a
subset of them for conversion to integrate-and-fire (IF) activations with
optimal bias shifts. Additionally, we derive the leaky-integrate-and-fire (LIF)
activation functions converted from their non-spiking LSTM counterparts which
justifies the need to jointly optimize the weights, threshold, and leak
parameter. We also propose a pipelined parallel processing scheme which hides
the SNN time steps, significantly improving system latency, especially for long
sequences. The resulting SNNs have high activation sparsity and require only
accumulate operations (AC), in contrast to expensive multiply-and-accumulates
(MAC) needed for ANNs, except for the input layer when using direct encoding,
yielding significant improvements in energy efficiency. We evaluate our
framework on sequential learning tasks including temporal MNIST, Google Speech
Commands (GSC), and UCI Smartphone datasets on different LSTM architectures. We
obtain test accuracy of 94.75% with only 2 time steps with direct encoding on
the GSC dataset with 4.1x lower energy than an iso-architecture standard LSTM.
Related papers
- Towards Low-latency Event-based Visual Recognition with Hybrid Step-wise Distillation Spiking Neural Networks [50.32980443749865]
Spiking neural networks (SNNs) have garnered significant attention for their low power consumption and high biologicalability.
Current SNNs struggle to balance accuracy and latency in neuromorphic datasets.
We propose Step-wise Distillation (HSD) method, tailored for neuromorphic datasets.
arXiv Detail & Related papers (2024-09-19T06:52:34Z) - Unlocking the Power of LSTM for Long Term Time Series Forecasting [27.245021350821638]
We propose a simple yet efficient algorithm named P-sLSTM built upon sLSTM by incorporating patching and channel independence.
These modifications substantially enhance sLSTM's performance in TSF, achieving state-of-the-art results.
arXiv Detail & Related papers (2024-08-19T13:59:26Z) - Spiking Neural Networks with Dynamic Time Steps for Vision Transformers [10.118436208925413]
Spiking Neural Networks (SNNs) have emerged as a popular-temporal computing paradigm vision for complex tasks.
We propose a novel training framework that dynamically allocates the number of time steps to each ViT module depending on a trainable score.
We evaluate our training framework and resulting SNNs on image recognition tasks including CIFAR10, CIFAR100, and ImageNet with different ViT architectures.
arXiv Detail & Related papers (2023-11-28T03:30:43Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Image Classification using Sequence of Pixels [3.04585143845864]
This study compares sequential image classification methods based on recurrent neural networks.
We describe methods based on Long-Short-Term memory(LSTM), bidirectional Long-Short-Term memory(BiLSTM) architectures, etc.
arXiv Detail & Related papers (2022-09-23T09:42:44Z) - Hybrid SNN-ANN: Energy-Efficient Classification and Object Detection for
Event-Based Vision [64.71260357476602]
Event-based vision sensors encode local pixel-wise brightness changes in streams of events rather than image frames.
Recent progress in object recognition from event-based sensors has come from conversions of deep neural networks.
We propose a hybrid architecture for end-to-end training of deep neural networks for event-based pattern recognition and object detection.
arXiv Detail & Related papers (2021-12-06T23:45:58Z) - A journey in ESN and LSTM visualisations on a language task [77.34726150561087]
We trained ESNs and LSTMs on a Cross-Situationnal Learning (CSL) task.
The results are of three kinds: performance comparison, internal dynamics analyses and visualization of latent space.
arXiv Detail & Related papers (2020-12-03T08:32:01Z) - LIAF-Net: Leaky Integrate and Analog Fire Network for Lightweight and
Efficient Spatiotemporal Information Processing [16.446511505488633]
Deep network LIAF-Net is built on LIF-SNN for efficienttemporal processing.
As atemporal layer, LIAF can also be used with traditional artificial neural network (ANN) layers jointly.
arXiv Detail & Related papers (2020-11-12T03:04:21Z) - Automatic Remaining Useful Life Estimation Framework with Embedded
Convolutional LSTM as the Backbone [5.927250637620123]
We propose a new LSTM variant called embedded convolutional LSTM (E NeuralTM)
In ETM a group of different 1D convolutions is embedded into the LSTM structure. Through this, the temporal information is preserved between and within windows.
We show the superiority of our proposed ETM approach over the state-of-the-art approaches on several widely used benchmark data sets for RUL Estimation.
arXiv Detail & Related papers (2020-08-10T08:34:20Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Object Tracking through Residual and Dense LSTMs [67.98948222599849]
Deep learning-based trackers based on LSTMs (Long Short-Term Memory) recurrent neural networks have emerged as a powerful alternative.
DenseLSTMs outperform Residual and regular LSTM, and offer a higher resilience to nuisances.
Our case study supports the adoption of residual-based RNNs for enhancing the robustness of other trackers.
arXiv Detail & Related papers (2020-06-22T08:20:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.