Probabilistic Modeling: Proving the Lottery Ticket Hypothesis in Spiking
Neural Network
- URL: http://arxiv.org/abs/2305.12148v1
- Date: Sat, 20 May 2023 09:27:34 GMT
- Title: Probabilistic Modeling: Proving the Lottery Ticket Hypothesis in Spiking
Neural Network
- Authors: Man Yao, Yuhong Chou, Guangshe Zhao, Xiawu Zheng, Yonghong Tian, Bo
Xu, Guoqi Li
- Abstract summary: Lottery Ticket Hypothesis (LTH) states that a randomly-d large neural network contains a small sub-network.
LTH opens up a new path for pruning network.
- Score: 30.924449325020767
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Lottery Ticket Hypothesis (LTH) states that a randomly-initialized large
neural network contains a small sub-network (i.e., winning tickets) which, when
trained in isolation, can achieve comparable performance to the large network.
LTH opens up a new path for network pruning. Existing proofs of LTH in
Artificial Neural Networks (ANNs) are based on continuous activation functions,
such as ReLU, which satisfying the Lipschitz condition. However, these
theoretical methods are not applicable in Spiking Neural Networks (SNNs) due to
the discontinuous of spiking function. We argue that it is possible to extend
the scope of LTH by eliminating Lipschitz condition. Specifically, we propose a
novel probabilistic modeling approach for spiking neurons with complicated
spatio-temporal dynamics. Then we theoretically and experimentally prove that
LTH holds in SNNs. According to our theorem, we conclude that pruning directly
in accordance with the weight size in existing SNNs is clearly not optimal. We
further design a new criterion for pruning based on our theory, which achieves
better pruning results than baseline.
Related papers
- Polynomially Over-Parameterized Convolutional Neural Networks Contain
Structured Strong Winning Lottery Tickets [4.020829863982153]
We prove the existence of structured Neuralworks that can approximate any sufficiently smaller network.
This result provides the first sub-exponential bound around the Strong Lottery Ticket Hypothesis.
arXiv Detail & Related papers (2023-11-16T12:38:45Z) - Benign Overfitting in Deep Neural Networks under Lazy Training [72.28294823115502]
We show that when the data distribution is well-separated, DNNs can achieve Bayes-optimal test error for classification.
Our results indicate that interpolating with smoother functions leads to better generalization.
arXiv Detail & Related papers (2023-05-30T19:37:44Z) - Strong Lottery Ticket Hypothesis with $\varepsilon$--perturbation [11.38723572165938]
We extend the theoretical guarantee of the strong Lottery Ticket Hypothesis to a scenario more similar to the original LTH.
By allowing an $varepsilon$-scale perturbation on the random initial weights, can we reduce the over- parameterization requirement for the candidate network in the strong LTH?
We show via experiments that the perturbed weight achieved by the projected SGD shows better performance under the strong LTH pruning.
arXiv Detail & Related papers (2022-10-29T12:22:17Z) - Robust Training and Verification of Implicit Neural Networks: A
Non-Euclidean Contractive Approach [64.23331120621118]
This paper proposes a theoretical and computational framework for training and robustness verification of implicit neural networks.
We introduce a related embedded network and show that the embedded network can be used to provide an $ell_infty$-norm box over-approximation of the reachable sets of the original network.
We apply our algorithms to train implicit neural networks on the MNIST dataset and compare the robustness of our models with the models trained via existing approaches in the literature.
arXiv Detail & Related papers (2022-08-08T03:13:24Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - On the Neural Tangent Kernel Analysis of Randomly Pruned Neural Networks [91.3755431537592]
We study how random pruning of the weights affects a neural network's neural kernel (NTK)
In particular, this work establishes an equivalence of the NTKs between a fully-connected neural network and its randomly pruned version.
arXiv Detail & Related papers (2022-03-27T15:22:19Z) - Dual Lottery Ticket Hypothesis [71.95937879869334]
Lottery Ticket Hypothesis (LTH) provides a novel view to investigate sparse network training and maintain its capacity.
In this work, we regard the winning ticket from LTH as the subnetwork which is in trainable condition and its performance as our benchmark.
We propose a simple sparse network training strategy, Random Sparse Network Transformation (RST), to substantiate our DLTH.
arXiv Detail & Related papers (2022-03-08T18:06:26Z) - Why Lottery Ticket Wins? A Theoretical Perspective of Sample Complexity
on Pruned Neural Networks [79.74580058178594]
We analyze the performance of training a pruned neural network by analyzing the geometric structure of the objective function.
We show that the convex region near a desirable model with guaranteed generalization enlarges as the neural network model is pruned.
arXiv Detail & Related papers (2021-10-12T01:11:07Z) - Random Vector Functional Link Networks for Function Approximation on Manifolds [8.535815777849786]
We show that single layer neural-networks with random input-to-hidden layer weights and biases have seen success in practice.
We further adapt this randomized neural network architecture to approximate functions on smooth, compact submanifolds of Euclidean space.
arXiv Detail & Related papers (2020-07-30T23:50:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.