Pursing the Sparse Limitation of Spiking Deep Learning Structures
- URL: http://arxiv.org/abs/2311.12060v1
- Date: Sat, 18 Nov 2023 17:00:40 GMT
- Title: Pursing the Sparse Limitation of Spiking Deep Learning Structures
- Authors: Hao Cheng, Jiahang Cao, Erjia Xiao, Mengshu Sun, Le Yang, Jize Zhang,
Xue Lin, Bhavya Kailkhura, Kaidi Xu, Renjing Xu
- Abstract summary: Spiking Neural Networks (SNNs) are garnering increased attention for their superior computation and energy efficiency.
We introduce an innovative algorithm capable of simultaneously identifying both weight and patch-level winning tickets.
We demonstrate that our spiking lottery ticket achieves comparable or superior performance even when the model structure is extremely sparse.
- Score: 42.334835610250714
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Spiking Neural Networks (SNNs), a novel brain-inspired algorithm, are
garnering increased attention for their superior computation and energy
efficiency over traditional artificial neural networks (ANNs). To facilitate
deployment on memory-constrained devices, numerous studies have explored SNN
pruning. However, these efforts are hindered by challenges such as scalability
challenges in more complex architectures and accuracy degradation. Amidst these
challenges, the Lottery Ticket Hypothesis (LTH) emerges as a promising pruning
strategy. It posits that within dense neural networks, there exist winning
tickets or subnetworks that are sparser but do not compromise performance. To
explore a more structure-sparse and energy-saving model, we investigate the
unique synergy of SNNs with LTH and design two novel spiking winning tickets to
push the boundaries of sparsity within SNNs. Furthermore, we introduce an
innovative algorithm capable of simultaneously identifying both weight and
patch-level winning tickets, enabling the achievement of sparser structures
without compromising on the final model's performance. Through comprehensive
experiments on both RGB-based and event-based datasets, we demonstrate that our
spiking lottery ticket achieves comparable or superior performance even when
the model structure is extremely sparse.
Related papers
- Parallel Hyperparameter Optimization Of Spiking Neural Network [0.5371337604556311]
Spiking Neural Networks (SNNs) are based on a more biologically inspired approach than usual artificial neural networks.
We tackle the signal loss issue of SNNs to what we call silent networks.
By defining an early stopping criterion, we were able to instantiate larger and more flexible search spaces.
arXiv Detail & Related papers (2024-03-01T11:11:59Z) - Multicoated and Folded Graph Neural Networks with Strong Lottery Tickets [3.0894823679470087]
This paper introduces the Multi-Stage Folding and Unshared Masks methods to expand the search space in terms of both architecture and parameters.
By achieving high sparsity, competitive performance, and high memory efficiency with up to 98.7% reduction, it demonstrates suitability for energy-efficient graph processing.
arXiv Detail & Related papers (2023-12-06T02:16:44Z) - Gaining the Sparse Rewards by Exploring Lottery Tickets in Spiking Neural Network [8.210103222339784]
Spiking Neural Networks (SNNs) offer a promising solution due to their low-latency and low-energy properties over traditional Artificial Neural Networks (ANNs)
This paper delves into the spiking-based LTs (SLTs), examining their unique properties and potential for extreme efficiency.
A sparse algorithm tailored for spiking transformer structure, which incorporates convolution operations into the Patch Embedding Projection (ConvPEP) module, has been proposed to achieve Multi-level Sparsity (MultiSp)
arXiv Detail & Related papers (2023-09-23T08:24:36Z) - An Automata-Theoretic Approach to Synthesizing Binarized Neural Networks [13.271286153792058]
Quantized neural networks (QNNs) have been developed, with binarized neural networks (BNNs) restricted to binary values as a special case.
This paper presents an automata-theoretic approach to synthesizing BNNs that meet designated properties.
arXiv Detail & Related papers (2023-07-29T06:27:28Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Training High-Performance Low-Latency Spiking Neural Networks by
Differentiation on Spike Representation [70.75043144299168]
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware.
It is a challenge to efficiently train SNNs due to their non-differentiability.
We propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance.
arXiv Detail & Related papers (2022-05-01T12:44:49Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Hybrid SNN-ANN: Energy-Efficient Classification and Object Detection for
Event-Based Vision [64.71260357476602]
Event-based vision sensors encode local pixel-wise brightness changes in streams of events rather than image frames.
Recent progress in object recognition from event-based sensors has come from conversions of deep neural networks.
We propose a hybrid architecture for end-to-end training of deep neural networks for event-based pattern recognition and object detection.
arXiv Detail & Related papers (2021-12-06T23:45:58Z) - SpikeMS: Deep Spiking Neural Network for Motion Segmentation [7.491944503744111]
textitSpikeMS is the first deep encoder-decoder SNN architecture for the real-world large-scale problem of motion segmentation.
We show that textitSpikeMS is capable of textitincremental predictions, or predictions from smaller amounts of test data than it is trained on.
arXiv Detail & Related papers (2021-05-13T21:34:55Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.