Hessian-Aware Pruning and Optimal Neural Implant
- URL: http://arxiv.org/abs/2101.08940v2
- Date: Sat, 6 Feb 2021 21:11:04 GMT
- Title: Hessian-Aware Pruning and Optimal Neural Implant
- Authors: Shixing Yu, Zhewei Yao, Amir Gholami, Zhen Dong, Michael W Mahoney,
Kurt Keutzer
- Abstract summary: Pruning is an effective method to reduce the memory footprint and FLOPs associated with neural network models.
We introduce a new Hessian Aware Pruning method coupled with a Neural Implant approach that uses second-order sensitivity as a metric for structured pruning.
- Score: 74.3282611517773
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pruning is an effective method to reduce the memory footprint and FLOPs
associated with neural network models. However, existing structured-pruning
methods often result in significant accuracy degradation for moderate pruning
levels. To address this problem, we introduce a new Hessian Aware Pruning (HAP)
method coupled with a Neural Implant approach that uses second-order
sensitivity as a metric for structured pruning. The basic idea is to prune
insensitive components and to use a Neural Implant for moderately sensitive
components, instead of completely pruning them. For the latter approach, the
moderately sensitive components are replaced with with a low rank implant that
is smaller and less computationally expensive than the original component. We
use the relative Hessian trace to measure sensitivity, as opposed to the
magnitude based sensitivity metric commonly used in the literature. We test HAP
on multiple models on CIFAR-10/ImageNet, and we achieve new state-of-the-art
results. Specifically, HAP achieves 94.3\% accuracy ($<0.1\%$ degradation) on
PreResNet29 (CIFAR-10), with more than 70\% of parameters pruned. Moreover, for
ResNet50 HAP achieves 75.1\% top-1 accuracy (0.5\% degradation) on ImageNet,
after pruning more than half of the parameters. The framework has been open
sourced and available online.
Related papers
- DRIVE: Dual Gradient-Based Rapid Iterative Pruning [2.209921757303168]
Modern deep neural networks (DNNs) consist of millions of parameters, necessitating high-performance computing during training and inference.
Traditional pruning methods that are applied post-training focus on streamlining inference, but there are recent efforts to leverage sparsity early on by pruning before training.
We present Dual Gradient-Based Rapid Iterative Pruning (DRIVE), which leverages dense training for initial epochs to counteract the randomness inherent at the inception.
arXiv Detail & Related papers (2024-04-01T20:44:28Z) - Instant Complexity Reduction in CNNs using Locality-Sensitive Hashing [50.79602839359522]
We propose HASTE (Hashing for Tractable Efficiency), a parameter-free and data-free module that acts as a plug-and-play replacement for any regular convolution module.
We are able to drastically compress latent feature maps without sacrificing much accuracy by using locality-sensitive hashing (LSH)
In particular, we are able to instantly drop 46.72% of FLOPs while only losing 1.25% accuracy by just swapping the convolution modules in a ResNet34 on CIFAR-10 for our HASTE module.
arXiv Detail & Related papers (2023-09-29T13:09:40Z) - Sr-init: An interpretable layer pruning method [11.184351630458265]
We propose a novel layer pruning method by exploring the Re-initialization.
Our SR-init method is inspired by the discovery that the accuracy drop due to re-initialization differs in various layers.
We experimentally verify the interpretability of SR-init via feature visualization.
arXiv Detail & Related papers (2023-03-14T07:26:55Z) - WeightMom: Learning Sparse Networks using Iterative Momentum-based
pruning [0.0]
We propose a weight based pruning approach in which the weights are pruned gradually based on their momentum of the previous iterations.
We evaluate our approach on networks such as AlexNet, VGG16 and ResNet50 with image classification datasets such as CIFAR-10 and CIFAR-100.
arXiv Detail & Related papers (2022-08-11T07:13:59Z) - Can pruning improve certified robustness of neural networks? [106.03070538582222]
We show that neural network pruning can improve empirical robustness of deep neural networks (NNs)
Our experiments show that by appropriately pruning an NN, its certified accuracy can be boosted up to 8.2% under standard training.
We additionally observe the existence of certified lottery tickets that can match both standard and certified robust accuracies of the original dense models.
arXiv Detail & Related papers (2022-06-15T05:48:51Z) - Pruning with Compensation: Efficient Channel Pruning for Deep
Convolutional Neural Networks [0.9712140341805068]
A highly efficient pruning method is proposed to significantly reduce the cost of pruning DCNN.
Our method shows competitive pruning performance among the state-of-the-art retraining-based pruning methods.
arXiv Detail & Related papers (2021-08-31T10:17:36Z) - Effective Model Sparsification by Scheduled Grow-and-Prune Methods [73.03533268740605]
We propose a novel scheduled grow-and-prune (GaP) methodology without pre-training the dense models.
Experiments have shown that such models can match or beat the quality of highly optimized dense models at 80% sparsity on a variety of tasks.
arXiv Detail & Related papers (2021-06-18T01:03:13Z) - Toward Compact Deep Neural Networks via Energy-Aware Pruning [2.578242050187029]
We propose a novel energy-aware pruning method that quantifies the importance of each filter in the network using nuclear-norm (NN)
We achieve competitive results with 40.4/49.8% of FLOPs and 45.9/52.9% of parameter reduction with 94.13/94.61% in the Top-1 accuracy with ResNet-56/110 on CIFAR-10.
arXiv Detail & Related papers (2021-03-19T15:33:16Z) - Non-Parametric Adaptive Network Pruning [125.4414216272874]
We introduce non-parametric modeling to simplify the algorithm design.
Inspired by the face recognition community, we use a message passing algorithm to obtain an adaptive number of exemplars.
EPruner breaks the dependency on the training data in determining the "important" filters.
arXiv Detail & Related papers (2021-01-20T06:18:38Z) - Neural Pruning via Growing Regularization [82.9322109208353]
We extend regularization to tackle two central problems of pruning: pruning schedule and weight importance scoring.
Specifically, we propose an L2 regularization variant with rising penalty factors and show it can bring significant accuracy gains.
The proposed algorithms are easy to implement and scalable to large datasets and networks in both structured and unstructured pruning.
arXiv Detail & Related papers (2020-12-16T20:16:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.