Advancing Deep Residual Learning by Solving the Crux of Degradation in
Spiking Neural Networks
- URL: http://arxiv.org/abs/2201.07209v2
- Date: Thu, 17 Feb 2022 07:37:21 GMT
- Title: Advancing Deep Residual Learning by Solving the Crux of Degradation in
Spiking Neural Networks
- Authors: Yifan Hu, Yujie Wu, Lei Deng, Guoqi Li
- Abstract summary: Residual learning and shortcuts have been evidenced as an important approach for training deep neural networks.
This paper proposes a novel residual block for SNNs, which is able to significantly extend the depth of directly trained SNNs.
- Score: 21.26300397341615
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Despite the rapid progress of neuromorphic computing, the inadequate depth
and the resulting insufficient representation power of spiking neural networks
(SNNs) severely restrict their application scope in practice. Residual learning
and shortcuts have been evidenced as an important approach for training deep
neural networks, but rarely did previous work assess their applicability to the
characteristics of spike-based communication and spatiotemporal dynamics. This
negligence leads to impeded information flow and the accompanying degradation
problem. In this paper, we identify the crux and then propose a novel residual
block for SNNs, which is able to significantly extend the depth of directly
trained SNNs, e.g., up to 482 layers on CIFAR-10 and 104 layers on ImageNet,
without observing any slight degradation problem. We validate the effectiveness
of our methods on both frame-based and neuromorphic datasets, and our
SRM-ResNet104 achieves a superior result of 76.02% accuracy on ImageNet, the
first time in the domain of directly trained SNNs. The great energy efficiency
is estimated and the resulting networks need on average only one spike per
neuron for classifying an input sample. We believe our powerful and scalable
modeling will provide a strong support for further exploration of SNNs.
Related papers
- STOP: Spatiotemporal Orthogonal Propagation for Weight-Threshold-Leakage Synergistic Training of Deep Spiking Neural Networks [11.85044871205734]
Deep neural network (SNN) models based on sparsely sparse binary activations lack efficient and high-accuracy SNN deep learning algorithms.
Our algorithm enables fully synergistic learning algorithm firing synaptic weights as well as thresholds and spiking factors in neurons to improve SNN accuracy.
Under a unified temporally-forward trace-based framework, we mitigate the huge memory requirement for storing neural states of all time-steps in the forward pass.
Our method is more plausible for edge intelligent scenarios where resources are limited but high-accuracy in-situ learning is desired.
arXiv Detail & Related papers (2024-11-17T14:15:54Z) - Addressing caveats of neural persistence with deep graph persistence [54.424983583720675]
We find that the variance of network weights and spatial concentration of large weights are the main factors that impact neural persistence.
We propose an extension of the filtration underlying neural persistence to the whole neural network instead of single layers.
This yields our deep graph persistence measure, which implicitly incorporates persistent paths through the network and alleviates variance-related issues.
arXiv Detail & Related papers (2023-07-20T13:34:11Z) - PC-SNN: Supervised Learning with Local Hebbian Synaptic Plasticity based
on Predictive Coding in Spiking Neural Networks [1.6172800007896282]
We propose a novel learning algorithm inspired by predictive coding theory.
We show that it can perform supervised learning fully autonomously and successfully as the backprop.
This method achieves a favorable performance compared to the state-of-the-art multi-layer SNNs.
arXiv Detail & Related papers (2022-11-24T09:56:02Z) - Multi-Level Firing with Spiking DS-ResNet: Enabling Better and Deeper
Directly-Trained Spiking Neural Networks [19.490903216456758]
Spiking neural networks (SNNs) are neural networks with asynchronous discrete and sparse characteristics.
We propose a multi-level firing (MLF) method based on the existing spiking-suppressed residual network (spiking DS-ResNet)
arXiv Detail & Related papers (2022-10-12T16:39:46Z) - Training High-Performance Low-Latency Spiking Neural Networks by
Differentiation on Spike Representation [70.75043144299168]
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware.
It is a challenge to efficiently train SNNs due to their non-differentiability.
We propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance.
arXiv Detail & Related papers (2022-05-01T12:44:49Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Advancing Residual Learning towards Powerful Deep Spiking Neural
Networks [16.559670769601038]
Residual learning and shortcuts have been evidenced as an important approach for training deep neural networks.
MS-ResNet is able to significantly extend the depth of directly trained SNNs.
MS-ResNet 104 achieves 76.02% accuracy on ImageNet, the first time in the domain of directly trained SNNs.
arXiv Detail & Related papers (2021-12-15T05:47:21Z) - HIRE-SNN: Harnessing the Inherent Robustness of Energy-Efficient Deep
Spiking Neural Networks by Training with Crafted Input Noise [13.904091056365765]
We present an SNN training algorithm that uses crafted input noise and incurs no additional training time.
Compared to standard trained direct input SNNs, our trained models yield improved classification accuracy of up to 13.7%.
Our models also outperform inherently robust SNNs trained on rate-coded inputs with improved or similar classification performance on attack-generated images.
arXiv Detail & Related papers (2021-10-06T16:48:48Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.