EagleEye: Fast Sub-net Evaluation for Efficient Neural Network Pruning
- URL: http://arxiv.org/abs/2007.02491v2
- Date: Wed, 5 Aug 2020 09:32:58 GMT
- Title: EagleEye: Fast Sub-net Evaluation for Efficient Neural Network Pruning
- Authors: Bailin Li, Bowen Wu, Jiang Su, Guangrun Wang and Liang Lin
- Abstract summary: EagleEye is a simple yet efficient evaluation component based on adaptive batch normalization.
It unveils a strong correlation between different pruned structures and their final settled accuracy.
This module is also general to plug-in and improve some existing pruning algorithms.
- Score: 82.54669314604097
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Finding out the computational redundant part of a trained Deep Neural Network
(DNN) is the key question that pruning algorithms target on. Many algorithms
try to predict model performance of the pruned sub-nets by introducing various
evaluation methods. But they are either inaccurate or very complicated for
general application. In this work, we present a pruning method called EagleEye,
in which a simple yet efficient evaluation component based on adaptive batch
normalization is applied to unveil a strong correlation between different
pruned DNN structures and their final settled accuracy. This strong correlation
allows us to fast spot the pruned candidates with highest potential accuracy
without actually fine-tuning them. This module is also general to plug-in and
improve some existing pruning algorithms. EagleEye achieves better pruning
performance than all of the studied pruning algorithms in our experiments.
Concretely, to prune MobileNet V1 and ResNet-50, EagleEye outperforms all
compared methods by up to 3.8%. Even in the more challenging experiments of
pruning the compact model of MobileNet V1, EagleEye achieves the highest
accuracy of 70.9% with an overall 50% operations (FLOPs) pruned. All accuracy
results are Top-1 ImageNet classification accuracy. Source code and models are
accessible to open-source community
https://github.com/anonymous47823493/EagleEye .
Related papers
- Towards Better Out-of-Distribution Generalization of Neural Algorithmic
Reasoning Tasks [51.8723187709964]
We study the OOD generalization of neural algorithmic reasoning tasks.
The goal is to learn an algorithm from input-output pairs using deep neural networks.
arXiv Detail & Related papers (2022-11-01T18:33:20Z) - EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for
Mobile Vision Applications [68.35683849098105]
We introduce split depth-wise transpose attention (SDTA) encoder that splits input tensors into multiple channel groups.
Our EdgeNeXt model with 1.3M parameters achieves 71.2% top-1 accuracy on ImageNet-1K.
Our EdgeNeXt model with 5.6M parameters achieves 79.4% top-1 accuracy on ImageNet-1K.
arXiv Detail & Related papers (2022-06-21T17:59:56Z) - Automatic Neural Network Pruning that Efficiently Preserves the Model
Accuracy [2.538209532048867]
pruning filters is a common solution, but most existing pruning methods do not preserve the model accuracy efficiently.
We propose an automatic pruning method that learns which neurons to preserve in order to maintain the model accuracy while reducing the FLOPs to a predefined target.
We achieve a 52.00% FLOPs reduction on ResNet-50, with a Top-1 accuracy of 47.51% after pruning and a state-of-the-art (SOTA) accuracy of 76.63% after finetuning.
arXiv Detail & Related papers (2021-11-18T11:29:35Z) - Neural network relief: a pruning algorithm based on neural activity [47.57448823030151]
We propose a simple importance-score metric that deactivates unimportant connections.
We achieve comparable performance for LeNet architectures on MNIST.
The algorithm is not designed to minimize FLOPs when considering current hardware and software implementations.
arXiv Detail & Related papers (2021-09-22T15:33:49Z) - AIP: Adversarial Iterative Pruning Based on Knowledge Transfer for
Convolutional Neural Networks [7.147985297123097]
convolutional neural networks (CNNs) take a fair amount of computation cost.
Current pruning methods can compress CNNs with little performance drop, but when the pruning ratio increases, the accuracy loss is more serious.
We propose a novel adversarial iterative pruning method (AIP) for CNNs based on knowledge transfer.
arXiv Detail & Related papers (2021-08-31T02:38:36Z) - Effective Model Sparsification by Scheduled Grow-and-Prune Methods [73.03533268740605]
We propose a novel scheduled grow-and-prune (GaP) methodology without pre-training the dense models.
Experiments have shown that such models can match or beat the quality of highly optimized dense models at 80% sparsity on a variety of tasks.
arXiv Detail & Related papers (2021-06-18T01:03:13Z) - ACP: Automatic Channel Pruning via Clustering and Swarm Intelligence
Optimization for CNN [6.662639002101124]
convolutional neural network (CNN) gets deeper and wider in recent years.
Existing magnitude-based pruning methods are efficient, but the performance of the compressed network is unpredictable.
We propose a novel automatic channel pruning method (ACP)
ACP is evaluated against several state-of-the-art CNNs on three different classification datasets.
arXiv Detail & Related papers (2021-01-16T08:56:38Z) - Holistic Filter Pruning for Efficient Deep Neural Networks [25.328005340524825]
"Holistic Filter Pruning" (HFP) is a novel approach for common DNN training that is easy to implement and enables to specify accurate pruning rates.
In various experiments, we give insights into the training and achieve state-of-the-art performance on CIFAR-10 and ImageNet.
arXiv Detail & Related papers (2020-09-17T09:23:36Z) - AdderNet: Do We Really Need Multiplications in Deep Learning? [159.174891462064]
We present adder networks (AdderNets) to trade massive multiplications in deep neural networks for much cheaper additions to reduce computation costs.
We develop a special back-propagation approach for AdderNets by investigating the full-precision gradient.
As a result, the proposed AdderNets can achieve 74.9% Top-1 accuracy 91.7% Top-5 accuracy using ResNet-50 on the ImageNet dataset.
arXiv Detail & Related papers (2019-12-31T06:56:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.