Cyclical Pruning for Sparse Neural Networks
- URL: http://arxiv.org/abs/2202.01290v1
- Date: Wed, 2 Feb 2022 21:29:07 GMT
- Title: Cyclical Pruning for Sparse Neural Networks
- Authors: Suraj Srinivas, Andrey Kuzmin, Markus Nagel, Mart van Baalen, Andrii
Skliar, Tijmen Blankevoort
- Abstract summary: Current methods for pruning neural network weights iteratively apply magnitude-based pruning on the model weights and re-train the resulting model to recover lost accuracy.
We show that such strategies do not allow for the recovery of erroneously pruned weights.
We propose a simple strategy called textitcyclical pruning which requires the pruning schedule to be periodic and allows for weights pruned erroneously in one cycle to recover in subsequent ones.
- Score: 21.13692106042014
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Current methods for pruning neural network weights iteratively apply
magnitude-based pruning on the model weights and re-train the resulting model
to recover lost accuracy. In this work, we show that such strategies do not
allow for the recovery of erroneously pruned weights. To enable weight
recovery, we propose a simple strategy called \textit{cyclical pruning} which
requires the pruning schedule to be periodic and allows for weights pruned
erroneously in one cycle to recover in subsequent ones. Experimental results on
both linear models and large-scale deep neural networks show that cyclical
pruning outperforms existing pruning algorithms, especially at high sparsity
ratios. Our approach is easy to tune and can be readily incorporated into
existing pruning pipelines to boost performance.
Related papers
- A Recovery Guarantee for Sparse Neural Networks [55.65046795628564]
We prove the first guarantees of sparse recovery for ReLU neural networks.<n>We find performance that is competitive with, and often exceeds, a high-performing but memory-inefficient baseline.
arXiv Detail & Related papers (2025-09-24T17:10:48Z) - Cyclic Sparse Training: Is it Enough? [17.02564269659367]
We propose SCULPT-ing, i.e., repeated cyclic training of any sparse mask followed by a single pruning step to couple the parameters and the mask.
This is able to match the performance of state-of-the-art iterative pruning methods in the high sparsity regime at reduced computational cost.
arXiv Detail & Related papers (2024-06-04T20:40:27Z) - To prune or not to prune : A chaos-causality approach to principled
pruning of dense neural networks [1.9249287163937978]
We introduce the concept of chaos in learning (Lyapunov exponents) via weight updates and exploiting causality to identify the causal weights responsible for misclassification.
Such a pruned network maintains the original performance and retains feature explainability.
arXiv Detail & Related papers (2023-08-19T09:17:33Z) - Trainability Preserving Neural Structured Pruning [64.65659982877891]
We present trainability preserving pruning (TPP), a regularization-based structured pruning method that can effectively maintain trainability during sparsification.
TPP can compete with the ground-truth dynamical isometry recovery method on linear networks.
It delivers encouraging performance in comparison to many top-performing filter pruning methods.
arXiv Detail & Related papers (2022-07-25T21:15:47Z) - Can pruning improve certified robustness of neural networks? [106.03070538582222]
We show that neural network pruning can improve empirical robustness of deep neural networks (NNs)
Our experiments show that by appropriately pruning an NN, its certified accuracy can be boosted up to 8.2% under standard training.
We additionally observe the existence of certified lottery tickets that can match both standard and certified robust accuracies of the original dense models.
arXiv Detail & Related papers (2022-06-15T05:48:51Z) - Boosting Pruned Networks with Linear Over-parameterization [8.796518772724955]
Structured pruning compresses neural networks by reducing channels (filters) for fast inference and low footprint at run-time.
To restore accuracy after pruning, fine-tuning is usually applied to pruned networks.
We propose a novel method that first linearly over- parameterizes the compact layers in pruned networks to enlarge the number of fine-tuning parameters.
arXiv Detail & Related papers (2022-04-25T05:30:26Z) - Pruning with Compensation: Efficient Channel Pruning for Deep
Convolutional Neural Networks [0.9712140341805068]
A highly efficient pruning method is proposed to significantly reduce the cost of pruning DCNN.
Our method shows competitive pruning performance among the state-of-the-art retraining-based pruning methods.
arXiv Detail & Related papers (2021-08-31T10:17:36Z) - Sparse Training via Boosting Pruning Plasticity with Neuroregeneration [79.78184026678659]
We study the effect of pruning throughout training from the perspective of pruning plasticity.
We design a novel gradual magnitude pruning (GMP) method, named gradual pruning with zero-cost neuroregeneration (GraNet) and its dynamic sparse training (DST) variant (GraNet-ST)
Perhaps most impressively, the latter for the first time boosts the sparse-to-sparse training performance over various dense-to-sparse methods by a large margin with ResNet-50 on ImageNet.
arXiv Detail & Related papers (2021-06-19T02:09:25Z) - Spectral Pruning for Recurrent Neural Networks [0.0]
Pruning techniques for neural networks with a recurrent architecture, such as the recurrent neural network (RNN), are strongly desired for their application to edge-computing devices.
In this paper, we propose an appropriate pruning algorithm for RNNs inspired by "spectral pruning", and provide the generalization error bounds for compressed RNNs.
arXiv Detail & Related papers (2021-05-23T00:30:59Z) - Neural Pruning via Growing Regularization [82.9322109208353]
We extend regularization to tackle two central problems of pruning: pruning schedule and weight importance scoring.
Specifically, we propose an L2 regularization variant with rising penalty factors and show it can bring significant accuracy gains.
The proposed algorithms are easy to implement and scalable to large datasets and networks in both structured and unstructured pruning.
arXiv Detail & Related papers (2020-12-16T20:16:28Z) - Exploring Weight Importance and Hessian Bias in Model Pruning [55.75546858514194]
We provide a principled exploration of pruning by building on a natural notion of importance.
For linear models, we show that this notion of importance is captured by scaling which connects to the well-known Hessian-based pruning algorithm.
We identify settings in which weights become more important despite becoming smaller, which in turn leads to a catastrophic failure of magnitude-based pruning.
arXiv Detail & Related papers (2020-06-19T00:15:55Z) - Compressive sensing with un-trained neural networks: Gradient descent
finds the smoothest approximation [60.80172153614544]
Un-trained convolutional neural networks have emerged as highly successful tools for image recovery and restoration.
We show that an un-trained convolutional neural network can approximately reconstruct signals and images that are sufficiently structured, from a near minimal number of random measurements.
arXiv Detail & Related papers (2020-05-07T15:57:25Z) - Lookahead: A Far-Sighted Alternative of Magnitude-based Pruning [83.99191569112682]
Magnitude-based pruning is one of the simplest methods for pruning neural networks.
We develop a simple pruning method, coined lookahead pruning, by extending the single layer optimization to a multi-layer optimization.
Our experimental results demonstrate that the proposed method consistently outperforms magnitude-based pruning on various networks.
arXiv Detail & Related papers (2020-02-12T05:38:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.