Sanity-Checking Pruning Methods: Random Tickets can Win the Jackpot
- URL: http://arxiv.org/abs/2009.11094v2
- Date: Thu, 22 Oct 2020 13:23:09 GMT
- Title: Sanity-Checking Pruning Methods: Random Tickets can Win the Jackpot
- Authors: Jingtong Su, Yihang Chen, Tianle Cai, Tianhao Wu, Ruiqi Gao, Liwei
Wang, Jason D. Lee
- Abstract summary: Conventional wisdom of pruning algorithms suggests that pruning methods exploit information from training data to find goodworks.
In this paper, we conduct sanity checks for the above beliefs on several recent unstructured pruning methods.
We propose a series of simple emphdata-independent prune ratios for each layer, and randomly prune each layer accordingly to get a subnetwork.
- Score: 55.37967301483917
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Network pruning is a method for reducing test-time computational resource
requirements with minimal performance degradation. Conventional wisdom of
pruning algorithms suggests that: (1) Pruning methods exploit information from
training data to find good subnetworks; (2) The architecture of the pruned
network is crucial for good performance. In this paper, we conduct sanity
checks for the above beliefs on several recent unstructured pruning methods and
surprisingly find that: (1) A set of methods which aims to find good
subnetworks of the randomly-initialized network (which we call "initial
tickets"), hardly exploits any information from the training data; (2) For the
pruned networks obtained by these methods, randomly changing the preserved
weights in each layer, while keeping the total number of preserved weights
unchanged per layer, does not affect the final performance. These findings
inspire us to choose a series of simple \emph{data-independent} prune ratios
for each layer, and randomly prune each layer accordingly to get a subnetwork
(which we call "random tickets"). Experimental results show that our zero-shot
random tickets outperform or attain a similar performance compared to existing
"initial tickets". In addition, we identify one existing pruning method that
passes our sanity checks. We hybridize the ratios in our random ticket with
this method and propose a new method called "hybrid tickets", which achieves
further improvement. (Our code is publicly available at
https://github.com/JingtongSu/sanity-checking-pruning)
Related papers
- Exploring the Performance of Pruning Methods in Neural Networks: An
Empirical Study of the Lottery Ticket Hypothesis [0.0]
We compare L1 unstructured pruning, Fisher pruning, and random pruning on different network architectures and pruning scenarios.
We propose and evaluate a new method for efficient computation of Fisher pruning, known as batched Fisher pruning.
arXiv Detail & Related papers (2023-03-26T21:46:34Z) - Revisiting Random Channel Pruning for Neural Network Compression [159.99002793644163]
Channel (or 3D filter) pruning serves as an effective way to accelerate the inference of neural networks.
In this paper, we try to determine the channel configuration of the pruned models by random search.
We show that this simple strategy works quite well compared with other channel pruning methods.
arXiv Detail & Related papers (2022-05-11T17:59:04Z) - Dual Lottery Ticket Hypothesis [71.95937879869334]
Lottery Ticket Hypothesis (LTH) provides a novel view to investigate sparse network training and maintain its capacity.
In this work, we regard the winning ticket from LTH as the subnetwork which is in trainable condition and its performance as our benchmark.
We propose a simple sparse network training strategy, Random Sparse Network Transformation (RST), to substantiate our DLTH.
arXiv Detail & Related papers (2022-03-08T18:06:26Z) - AdaPruner: Adaptive Channel Pruning and Effective Weights Inheritance [9.3421559369389]
We propose a pruning framework that adaptively determines the number of each layer's channels as well as the wights inheritance criteria for sub-network.
AdaPruner allows to obtain pruned network quickly, accurately and efficiently.
On ImageNet, we reduce 32.8% FLOPs of MobileNetV2 with only 0.62% decrease for top-1 accuracy, which exceeds all previous state-of-the-art channel pruning methods.
arXiv Detail & Related papers (2021-09-14T01:52:05Z) - FreeTickets: Accurate, Robust and Efficient Deep Ensemble by Training
with Dynamic Sparsity [74.58777701536668]
We introduce the FreeTickets concept, which can boost the performance of sparse convolutional neural networks over their dense network equivalents by a large margin.
We propose two novel efficient ensemble methods with dynamic sparsity, which yield in one shot many diverse and accurate tickets "for free" during the sparse training process.
arXiv Detail & Related papers (2021-06-28T10:48:20Z) - The Elastic Lottery Ticket Hypothesis [106.79387235014379]
Lottery Ticket Hypothesis raises keen attention to identifying sparse trainableworks or winning tickets.
The most effective method to identify such winning tickets is still Iterative Magnitude-based Pruning.
We propose a variety of strategies to tweak the winning tickets found from different networks of the same model family.
arXiv Detail & Related papers (2021-03-30T17:53:45Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Data-dependent Pruning to find the Winning Lottery Ticket [0.0]
Lottery Ticket Hypothesis postulates that a freshly neural network contains a small subnetwork that can be trained to achieve similar performance as the full network.
We conclude that incorporating a data dependent component into the pruning criterion consistently improves the performance of existing pruning algorithms.
arXiv Detail & Related papers (2020-06-25T12:48:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.