Juvenile state hypothesis: What we can learn from lottery ticket
hypothesis researches?
- URL: http://arxiv.org/abs/2109.03862v1
- Date: Wed, 8 Sep 2021 18:22:00 GMT
- Title: Juvenile state hypothesis: What we can learn from lottery ticket
hypothesis researches?
- Authors: Di Zhang
- Abstract summary: Original lottery ticket hypothesis performs pruning and weight resetting after training convergence.
We propose a strategy that combines the idea of neural network structure search with a pruning algorithm to alleviate this problem.
- Score: 1.701869491238765
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The proposition of lottery ticket hypothesis revealed the relationship
between network structure and initialization parameters and the learning
potential of neural networks. The original lottery ticket hypothesis performs
pruning and weight resetting after training convergence, exposing it to the
problem of forgotten learning knowledge and potential high cost of training.
Therefore, we propose a strategy that combines the idea of neural network
structure search with a pruning algorithm to alleviate this problem. This
algorithm searches and extends the network structure on existing winning ticket
sub-network to producing new winning ticket recursively. This allows the
training and pruning process to continue without compromising performance. A
new winning ticket sub-network with deeper network structure, better
generalization ability and better test performance can be obtained in this
recursive manner. This method can solve: the difficulty of training or
performance degradation of the sub-networks after pruning, the forgetting of
the weights of the original lottery ticket hypothesis and the difficulty of
generating winning ticket sub-network when the final network structure is not
given. We validate this strategy on the MNIST and CIFAR-10 datasets. And after
relating it to similar biological phenomena and relevant lottery ticket
hypothesis studies in recent years, we will further propose a new hypothesis to
discuss which factors that can keep a network juvenile, i.e., those possible
factors that influence the learning potential or generalization performance of
a neural network during training.
Related papers
- Dual Lottery Ticket Hypothesis [71.95937879869334]
Lottery Ticket Hypothesis (LTH) provides a novel view to investigate sparse network training and maintain its capacity.
In this work, we regard the winning ticket from LTH as the subnetwork which is in trainable condition and its performance as our benchmark.
We propose a simple sparse network training strategy, Random Sparse Network Transformation (RST), to substantiate our DLTH.
arXiv Detail & Related papers (2022-03-08T18:06:26Z) - Why Lottery Ticket Wins? A Theoretical Perspective of Sample Complexity
on Pruned Neural Networks [79.74580058178594]
We analyze the performance of training a pruned neural network by analyzing the geometric structure of the objective function.
We show that the convex region near a desirable model with guaranteed generalization enlarges as the neural network model is pruned.
arXiv Detail & Related papers (2021-10-12T01:11:07Z) - FreeTickets: Accurate, Robust and Efficient Deep Ensemble by Training
with Dynamic Sparsity [74.58777701536668]
We introduce the FreeTickets concept, which can boost the performance of sparse convolutional neural networks over their dense network equivalents by a large margin.
We propose two novel efficient ensemble methods with dynamic sparsity, which yield in one shot many diverse and accurate tickets "for free" during the sparse training process.
arXiv Detail & Related papers (2021-06-28T10:48:20Z) - The Elastic Lottery Ticket Hypothesis [106.79387235014379]
Lottery Ticket Hypothesis raises keen attention to identifying sparse trainableworks or winning tickets.
The most effective method to identify such winning tickets is still Iterative Magnitude-based Pruning.
We propose a variety of strategies to tweak the winning tickets found from different networks of the same model family.
arXiv Detail & Related papers (2021-03-30T17:53:45Z) - Good Students Play Big Lottery Better [84.6111281091602]
Lottery ticket hypothesis suggests that a dense neural network contains a sparse sub-network that can match the test accuracy of the original dense net.
Recent studies demonstrate that a sparse sub-network can still be obtained by using a rewinding technique.
This paper proposes a new, simpler and yet powerful technique for re-training the sub-network, called "Knowledge Distillation ticket" (KD ticket)
arXiv Detail & Related papers (2021-01-08T23:33:53Z) - Towards Practical Lottery Ticket Hypothesis for Adversarial Training [78.30684998080346]
We show there exists a subset of the aforementioned sub-networks that converge significantly faster during the training process.
As a practical application of our findings, we demonstrate that such sub-networks can help in cutting down the total time of adversarial training.
arXiv Detail & Related papers (2020-03-06T03:11:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.