Can We Find Strong Lottery Tickets in Generative Models?
- URL: http://arxiv.org/abs/2212.08311v1
- Date: Fri, 16 Dec 2022 07:20:28 GMT
- Title: Can We Find Strong Lottery Tickets in Generative Models?
- Authors: Sangyeop Yeo, Yoojin Jang, Jy-yong Sohn, Dongyoon Han, Jaejun Yoo
- Abstract summary: We find strong lottery tickets in generative models that achieve good generative performance without any weight update.
To the best of our knowledge, we are the first to show the existence of strong lottery tickets in generative models and provide an algorithm to find it.
- Score: 24.405555822170896
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Yes. In this paper, we investigate strong lottery tickets in generative
models, the subnetworks that achieve good generative performance without any
weight update. Neural network pruning is considered the main cornerstone of
model compression for reducing the costs of computation and memory.
Unfortunately, pruning a generative model has not been extensively explored,
and all existing pruning algorithms suffer from excessive weight-training
costs, performance degradation, limited generalizability, or complicated
training. To address these problems, we propose to find a strong lottery ticket
via moment-matching scores. Our experimental results show that the discovered
subnetwork can perform similarly or better than the trained dense model even
when only 10% of the weights remain. To the best of our knowledge, we are the
first to show the existence of strong lottery tickets in generative models and
provide an algorithm to find it stably. Our code and supplementary materials
are publicly available.
Related papers
- Finding Strong Lottery Ticket Networks with Genetic Algorithms [3.1267592104279776]
According to the Strong Lottery Ticket Hypothesis, every sufficiently large neural network with randomly weights contains a sub-network which already performs as well for a given task as the trained super-network.
We present the first approach based on a genetic algorithm to find such strong lottery ticket sub-networks without training or otherwise computing any gradient.
arXiv Detail & Related papers (2024-11-07T12:35:35Z) - Dual Lottery Ticket Hypothesis [71.95937879869334]
Lottery Ticket Hypothesis (LTH) provides a novel view to investigate sparse network training and maintain its capacity.
In this work, we regard the winning ticket from LTH as the subnetwork which is in trainable condition and its performance as our benchmark.
We propose a simple sparse network training strategy, Random Sparse Network Transformation (RST), to substantiate our DLTH.
arXiv Detail & Related papers (2022-03-08T18:06:26Z) - You are caught stealing my winning lottery ticket! Making a lottery
ticket claim its ownership [87.13642800792077]
Lottery ticket hypothesis (LTH) emerges as a promising framework to leverage a special sparse subnetwork.
Main resource bottleneck of LTH is however the extraordinary cost to find the sparse mask of the winning ticket.
Our setting adds a new dimension to the recently soaring interest in protecting against the intellectual property infringement of deep models.
arXiv Detail & Related papers (2021-10-30T03:38:38Z) - Lottery Jackpots Exist in Pre-trained Models [69.17690253938211]
We show that high-performing and sparse sub-networks without the involvement of weight training, termed "lottery jackpots", exist in pre-trained models with unexpanded width.
We propose a novel short restriction method to restrict change of masks that may have potential negative impacts on the training loss.
arXiv Detail & Related papers (2021-04-18T03:50:28Z) - The Elastic Lottery Ticket Hypothesis [106.79387235014379]
Lottery Ticket Hypothesis raises keen attention to identifying sparse trainableworks or winning tickets.
The most effective method to identify such winning tickets is still Iterative Magnitude-based Pruning.
We propose a variety of strategies to tweak the winning tickets found from different networks of the same model family.
arXiv Detail & Related papers (2021-03-30T17:53:45Z) - Lottery Ticket Implies Accuracy Degradation, Is It a Desirable
Phenomenon? [43.47794674403988]
In deep model compression, the recent finding "Lottery Ticket Hypothesis" (LTH) (Frankle & Carbin) pointed out that there could exist a winning ticket.
We investigate the underlying condition and rationale behind the winning property, and find that the underlying reason is largely attributed to the correlation between weights and final-trained weights.
We propose the "pruning & fine-tuning" method that consistently outperforms lottery ticket sparse training.
arXiv Detail & Related papers (2021-02-19T14:49:46Z) - Good Students Play Big Lottery Better [84.6111281091602]
Lottery ticket hypothesis suggests that a dense neural network contains a sparse sub-network that can match the test accuracy of the original dense net.
Recent studies demonstrate that a sparse sub-network can still be obtained by using a rewinding technique.
This paper proposes a new, simpler and yet powerful technique for re-training the sub-network, called "Knowledge Distillation ticket" (KD ticket)
arXiv Detail & Related papers (2021-01-08T23:33:53Z) - Winning Lottery Tickets in Deep Generative Models [64.79920299421255]
We show the existence of winning tickets in deep generative models such as GANs and VAEs.
We also demonstrate the transferability of winning tickets across different generative models.
arXiv Detail & Related papers (2020-10-05T21:45:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.