Efficient Training of Spiking Neural Networks with Temporally-Truncated
Local Backpropagation through Time
- URL: http://arxiv.org/abs/2201.07210v1
- Date: Mon, 13 Dec 2021 07:44:58 GMT
- Title: Efficient Training of Spiking Neural Networks with Temporally-Truncated
Local Backpropagation through Time
- Authors: Wenzhe Guo, Mohammed E. Fouda, Ahmed M. Eltawil, and Khaled Nabil
Salama
- Abstract summary: Training spiking neural networks (SNNs) has remained challenging due to complex neural dynamics and intrinsic non-differentiability in firing functions.
This work proposes an efficient and direct training algorithm for SNNs that integrates a locally-supervised training method with a temporally-truncated BPTT algorithm.
- Score: 1.926678651590519
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Directly training spiking neural networks (SNNs) has remained challenging due
to complex neural dynamics and intrinsic non-differentiability in firing
functions. The well-known backpropagation through time (BPTT) algorithm
proposed to train SNNs suffers from large memory footprint and prohibits
backward and update unlocking, making it impossible to exploit the potential of
locally-supervised training methods. This work proposes an efficient and direct
training algorithm for SNNs that integrates a locally-supervised training
method with a temporally-truncated BPTT algorithm. The proposed algorithm
explores both temporal and spatial locality in BPTT and contributes to
significant reduction in computational cost including GPU memory utilization,
main memory access and arithmetic operations. We thoroughly explore the design
space concerning temporal truncation length and local training block size and
benchmark their impact on classification accuracy of different networks running
different types of tasks. The results reveal that temporal truncation has a
negative effect on the accuracy of classifying frame-based datasets, but leads
to improvement in accuracy on dynamic-vision-sensor (DVS) recorded datasets. In
spite of resulting information loss, local training is capable of alleviating
overfitting. The combined effect of temporal truncation and local training can
lead to the slowdown of accuracy drop and even improvement in accuracy. In
addition, training deep SNNs models such as AlexNet classifying CIFAR10-DVS
dataset leads to 7.26% increase in accuracy, 89.94% reduction in GPU memory,
10.79% reduction in memory access, and 99.64% reduction in MAC operations
compared to the standard end-to-end BPTT.
Related papers
- Exploiting Symmetric Temporally Sparse BPTT for Efficient RNN Training [20.49255973077044]
This work describes a training algorithm for Delta RNNs that exploits temporal sparsity in the backward propagation phase to reduce computational requirements for training on the edge.
Results show a reduction of $sim$80% in matrix operations for training a 56k parameter Delta LSTM on the Fluent Speech Commands dataset with negligible accuracy loss.
We show that the proposed Delta RNN training will be useful for online incremental learning on edge devices with limited computing resources.
arXiv Detail & Related papers (2023-12-14T23:07:37Z) - Estimating Post-Synaptic Effects for Online Training of Feed-Forward
SNNs [0.27016900604393124]
Facilitating online learning in spiking neural networks (SNNs) is a key step in developing event-based models.
We propose Online Training with Postsynaptic Estimates (OTPE) for training feed-forward SNNs.
We show improved scaling for multi-layer networks using a novel approximation of temporal effects on the subsequent layer's activity.
arXiv Detail & Related papers (2023-11-07T16:53:39Z) - Towards Memory- and Time-Efficient Backpropagation for Training Spiking
Neural Networks [70.75043144299168]
Spiking Neural Networks (SNNs) are promising energy-efficient models for neuromorphic computing.
We propose the Spatial Learning Through Time (SLTT) method that can achieve high performance while greatly improving training efficiency.
Our method achieves state-of-the-art accuracy on ImageNet, while the memory cost and training time are reduced by more than 70% and 50%, respectively, compared with BPTT.
arXiv Detail & Related papers (2023-02-28T05:01:01Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Online Training Through Time for Spiking Neural Networks [66.7744060103562]
Spiking neural networks (SNNs) are promising brain-inspired energy-efficient models.
Recent progress in training methods has enabled successful deep SNNs on large-scale tasks with low latency.
We propose online training through time (OTTT) for SNNs, which is derived from BPTT to enable forward-in-time learning.
arXiv Detail & Related papers (2022-10-09T07:47:56Z) - Backpropagation with Biologically Plausible Spatio-Temporal Adjustment
For Training Deep Spiking Neural Networks [5.484391472233163]
The success of deep learning is inseparable from backpropagation.
We propose a biological plausible spatial adjustment, which rethinks the relationship between membrane potential and spikes.
Secondly, we propose a biologically plausible temporal adjustment making the error propagate across the spikes in the temporal dimension.
arXiv Detail & Related papers (2021-10-17T15:55:51Z) - FracTrain: Fractionally Squeezing Bit Savings Both Temporally and
Spatially for Efficient DNN Training [81.85361544720885]
We propose FracTrain that integrates progressive fractional quantization which gradually increases the precision of activations, weights, and gradients.
FracTrain reduces computational cost and hardware-quantified energy/latency of DNN training while achieving a comparable or better (-0.12%+1.87%) accuracy.
arXiv Detail & Related papers (2020-12-24T05:24:10Z) - Predicting Training Time Without Training [120.92623395389255]
We tackle the problem of predicting the number of optimization steps that a pre-trained deep network needs to converge to a given value of the loss function.
We leverage the fact that the training dynamics of a deep network during fine-tuning are well approximated by those of a linearized model.
We are able to predict the time it takes to fine-tune a model to a given loss without having to perform any training.
arXiv Detail & Related papers (2020-08-28T04:29:54Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.