SBPF: Sensitiveness Based Pruning Framework For Convolutional Neural
Network On Image Classification
- URL: http://arxiv.org/abs/2208.04588v1
- Date: Tue, 9 Aug 2022 08:05:19 GMT
- Title: SBPF: Sensitiveness Based Pruning Framework For Convolutional Neural
Network On Image Classification
- Authors: Yiheng Lu, Maoguo Gong, Wei Zhao, Kaiyuan Feng, and Hao Li
- Abstract summary: Pruning techniques are used to compress convolutional neural networks (CNNs) on image classification.
We propose a sensitiveness based method to evaluate the importance of each layer from the perspective of inference accuracy.
- Score: 23.271501988351268
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pruning techniques are used comprehensively to compress convolutional neural
networks (CNNs) on image classification. However, the majority of pruning
methods require a well pre-trained model to provide useful supporting
parameters, such as C1-norm, BatchNorm value and gradient information, which
may lead to inconsistency of filter evaluation if the parameters of the
pre-trained model are not well optimized. Therefore, we propose a sensitiveness
based method to evaluate the importance of each layer from the perspective of
inference accuracy by adding extra damage for the original model. Because the
performance of the accuracy is determined by the distribution of parameters
across all layers rather than individual parameter, the sensitiveness based
method will be robust to update of parameters. Namely, we can obtain similar
importance evaluation of each convolutional layer between the imperfect-trained
and fully trained models. For VGG-16 on CIFAR-10, even when the original model
is only trained with 50 epochs, we can get same evaluation of layer importance
as the results when the model is trained fully. Then we will remove filters
proportional from each layer by the quantified sensitiveness. Our sensitiveness
based pruning framework is verified efficiently on VGG-16, a customized Conv-4
and ResNet-18 with CIFAR-10, MNIST and CIFAR-100, respectively.
Related papers
- MPruner: Optimizing Neural Network Size with CKA-Based Mutual Information Pruning [7.262751938473306]
Pruning is a well-established technique that reduces the size of neural networks while mathematically guaranteeing accuracy preservation.
We develop a new pruning algorithm, MPruner, that leverages mutual information through vector similarity.
MPruner achieved up to a 50% reduction in parameters and memory usage for CNN and transformer-based models, with minimal to no loss in accuracy.
arXiv Detail & Related papers (2024-08-24T05:54:47Z) - Just How Flexible are Neural Networks in Practice? [89.80474583606242]
It is widely believed that a neural network can fit a training set containing at least as many samples as it has parameters.
In practice, however, we only find solutions via our training procedure, including the gradient and regularizers, limiting flexibility.
arXiv Detail & Related papers (2024-06-17T12:24:45Z) - Prompt Tuning for Parameter-efficient Medical Image Segmentation [79.09285179181225]
We propose and investigate several contributions to achieve a parameter-efficient but effective adaptation for semantic segmentation on two medical imaging datasets.
We pre-train this architecture with a dedicated dense self-supervision scheme based on assignments to online generated prototypes.
We demonstrate that the resulting neural network model is able to attenuate the gap between fully fine-tuned and parameter-efficiently adapted models.
arXiv Detail & Related papers (2022-11-16T21:55:05Z) - Pruning by Active Attention Manipulation [49.61707925611295]
Filter pruning of a CNN is typically achieved by applying discrete masks on the CNN's filter weights or activation maps, post-training.
Here, we present a new filter-importance-scoring concept named pruning by active attention manipulation (PAAM)
PAAM learns analog filter scores from the filter weights by optimizing a cost function regularized by an additive term in the scores.
arXiv Detail & Related papers (2022-10-20T09:17:02Z) - Entropy Induced Pruning Framework for Convolutional Neural Networks [30.89967076857665]
We propose a metric named Average Filter Information Entropy (AFIE) to measure the importance of each filter.
The proposed framework is able to yield a stable importance evaluation of each filter no matter whether the original model is trained fully.
arXiv Detail & Related papers (2022-08-13T14:35:08Z) - (Certified!!) Adversarial Robustness for Free! [116.6052628829344]
We certify 71% accuracy on ImageNet under adversarial perturbations constrained to be within a 2-norm of 0.5.
We obtain these results using only pretrained diffusion models and image classifiers, without requiring any fine tuning or retraining of model parameters.
arXiv Detail & Related papers (2022-06-21T17:27:27Z) - A Comprehensive Study of Image Classification Model Sensitivity to
Foregrounds, Backgrounds, and Visual Attributes [58.633364000258645]
We call this dataset RIVAL10 consisting of roughly $26k$ instances over $10$ classes.
We evaluate the sensitivity of a broad set of models to noise corruptions in foregrounds, backgrounds and attributes.
In our analysis, we consider diverse state-of-the-art architectures (ResNets, Transformers) and training procedures (CLIP, SimCLR, DeiT, Adversarial Training)
arXiv Detail & Related papers (2022-01-26T06:31:28Z) - Fire Together Wire Together: A Dynamic Pruning Approach with
Self-Supervised Mask Prediction [12.86325214182021]
Dynamic model pruning is a recent direction that allows for the inference of a different sub-network for each input sample during deployment.
Current dynamic methods rely on learning a continuous channel gating through regularization by inducing sparsity loss.
We show experiments on several neural architectures, such as VGG, ResNet, and MobileNet on CIFAR and ImageNet.
arXiv Detail & Related papers (2021-10-15T17:39:53Z) - Basis Scaling and Double Pruning for Efficient Inference in
Network-Based Transfer Learning [1.3467579878240454]
We decompose a convolutional layer into two layers: a convolutional layer with the orthonormal basis vectors as the filters, and a "BasisScalingConv" layer which is responsible for rescaling the features.
We can achieve pruning ratios up to 74.6% for CIFAR-10 and 98.9% for MNIST in model parameters.
arXiv Detail & Related papers (2021-08-06T00:04:02Z) - Non-Parametric Adaptive Network Pruning [125.4414216272874]
We introduce non-parametric modeling to simplify the algorithm design.
Inspired by the face recognition community, we use a message passing algorithm to obtain an adaptive number of exemplars.
EPruner breaks the dependency on the training data in determining the "important" filters.
arXiv Detail & Related papers (2021-01-20T06:18:38Z) - An Artificial Intelligence System for Combined Fruit Detection and
Georeferencing, Using RTK-Based Perspective Projection in Drone Imagery [0.0]
This work presents an Artificial Intelligence (AI) system, which detects and counts apples from aerial drone imagery of commercial orchards.
To reduce computational cost, a novel precursory stage to the network is designed to preprocess raw imagery into cropped images of individual trees.
Unique geospatial identifiers are allocated to these using the perspective projection model.
Experiments show that a k-means clustering approach, never before seen in literature for Faster R-CNN, resulted in the most significant improvements to calibrated mAP.
arXiv Detail & Related papers (2021-01-01T23:39:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.