ThinResNet: A New Baseline for Structured Convolutional Networks Pruning
- URL: http://arxiv.org/abs/2309.12854v1
- Date: Fri, 22 Sep 2023 13:28:18 GMT
- Title: ThinResNet: A New Baseline for Structured Convolutional Networks Pruning
- Authors: Hugo Tessier, Ghouti Boukli Hacene, Vincent Gripon
- Abstract summary: Pruning is a compression method which aims to improve the efficiency of neural networks by reducing their number of parameters.
In this work, we verify how results in the recent literature of pruning hold up against networks that underwent both state-of-the-art training methods and trivial model scaling.
- Score: 1.90298817989995
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Pruning is a compression method which aims to improve the efficiency of
neural networks by reducing their number of parameters while maintaining a good
performance, thus enhancing the performance-to-cost ratio in nontrivial ways.
Of particular interest are structured pruning techniques, in which whole
portions of parameters are removed altogether, resulting in easier to leverage
shrunk architectures. Since its growth in popularity in the recent years,
pruning gave birth to countless papers and contributions, resulting first in
critical inconsistencies in the way results are compared, and then to a
collective effort to establish standardized benchmarks. However, said
benchmarks are based on training practices that date from several years ago and
do not align with current practices. In this work, we verify how results in the
recent literature of pruning hold up against networks that underwent both
state-of-the-art training methods and trivial model scaling. We find that the
latter clearly and utterly outperform all the literature we compared to,
proving that updating standard pruning benchmarks and re-evaluating classical
methods in their light is an absolute necessity. We thus introduce a new
challenging baseline to compare structured pruning to: ThinResNet.
Related papers
- How to Prune Your Language Model: Recovering Accuracy on the "Sparsity
May Cry'' Benchmark [60.72725673114168]
We revisit the question of accurate BERT-pruning during fine-tuning on downstream datasets.
We propose a set of general guidelines for successful pruning, even on the challenging SMC benchmark.
arXiv Detail & Related papers (2023-12-21T03:11:30Z) - Accurate Neural Network Pruning Requires Rethinking Sparse Optimization [87.90654868505518]
We show the impact of high sparsity on model training using the standard computer vision and natural language processing sparsity benchmarks.
We provide new approaches for mitigating this issue for both sparse pre-training of vision models and sparse fine-tuning of language models.
arXiv Detail & Related papers (2023-08-03T21:49:14Z) - Pushing the Efficiency Limit Using Structured Sparse Convolutions [82.31130122200578]
We propose Structured Sparse Convolution (SSC), which leverages the inherent structure in images to reduce the parameters in the convolutional filter.
We show that SSC is a generalization of commonly used layers (depthwise, groupwise and pointwise convolution) in efficient architectures''
Architectures based on SSC achieve state-of-the-art performance compared to baselines on CIFAR-10, CIFAR-100, Tiny-ImageNet, and ImageNet classification benchmarks.
arXiv Detail & Related papers (2022-10-23T18:37:22Z) - Compact Neural Networks via Stacking Designed Basic Units [38.10212043168065]
This paper presents a new method termed TissueNet, which directly constructs compact neural networks with fewer weight parameters.
We formulate TissueNet in diverse popular backbones for comparison with the state-of-the-art pruning methods on different benchmark datasets.
Experiment results show that TissueNet can achieve comparable classification accuracy while saving up to around 80% FLOPs and 89.7% parameters.
arXiv Detail & Related papers (2022-05-03T14:04:49Z) - Back to Basics: Efficient Network Compression via IMP [22.586474627159287]
Iterative Magnitude Pruning (IMP) is one of the most established approaches for network pruning.
IMP is often argued that it reaches suboptimal states by not incorporating sparsification into the training phase.
We find that IMP with SLR for retraining can outperform state-of-the-art pruning-during-training approaches.
arXiv Detail & Related papers (2021-11-01T11:23:44Z) - Rethinking Network Pruning -- under the Pre-train and Fine-tune Paradigm [5.621336109915588]
We show for the first time that sparse pruning compresses a BERT model significantly more than reducing its number of channels and layers.
Our method outperforms the leading competitors with a 20-times weight/FLOPs compression and neglectable loss in prediction accuracy.
arXiv Detail & Related papers (2021-04-18T02:20:37Z) - Emerging Paradigms of Neural Network Pruning [82.9322109208353]
Pruning is adopted as a post-processing solution to this problem, which aims to remove unnecessary parameters in a neural network with little performance compromised.
Recent works challenge this belief by discovering random sparse networks which can be trained to match the performance with their dense counterpart.
This survey seeks to bridge the gap by proposing a general pruning framework so that the emerging pruning paradigms can be accommodated well with the traditional one.
arXiv Detail & Related papers (2021-03-11T05:01:52Z) - Manifold Regularized Dynamic Network Pruning [102.24146031250034]
This paper proposes a new paradigm that dynamically removes redundant filters by embedding the manifold information of all instances into the space of pruned networks.
The effectiveness of the proposed method is verified on several benchmarks, which shows better performance in terms of both accuracy and computational cost.
arXiv Detail & Related papers (2021-03-10T03:59:03Z) - Lost in Pruning: The Effects of Pruning Neural Networks beyond Test
Accuracy [42.15969584135412]
Neural network pruning is a popular technique used to reduce the inference costs of modern networks.
We evaluate whether the use of test accuracy alone in the terminating condition is sufficient to ensure that the resulting model performs well.
We find that pruned networks effectively approximate the unpruned model, however, the prune ratio at which pruned networks achieve commensurate performance varies significantly across tasks.
arXiv Detail & Related papers (2021-03-04T13:22:16Z) - Neural Pruning via Growing Regularization [82.9322109208353]
We extend regularization to tackle two central problems of pruning: pruning schedule and weight importance scoring.
Specifically, we propose an L2 regularization variant with rising penalty factors and show it can bring significant accuracy gains.
The proposed algorithms are easy to implement and scalable to large datasets and networks in both structured and unstructured pruning.
arXiv Detail & Related papers (2020-12-16T20:16:28Z) - Paying more attention to snapshots of Iterative Pruning: Improving Model
Compression via Ensemble Distillation [4.254099382808598]
Existing methods often iteratively prune networks to attain high compression ratio without incurring significant loss in performance.
We show that strong ensembles can be constructed from snapshots of iterative pruning, which achieve competitive performance and vary in network structure.
In standard image classification benchmarks such as CIFAR and Tiny-Imagenet, we advance state-of-the-art pruning ratio of structured pruning by integrating simple l1-norm filters pruning into our pipeline.
arXiv Detail & Related papers (2020-06-20T03:59:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.