Efficient CNNs via Passive Filter Pruning
- URL: http://arxiv.org/abs/2304.02319v1
- Date: Wed, 5 Apr 2023 09:19:19 GMT
- Title: Efficient CNNs via Passive Filter Pruning
- Authors: Arshdeep Singh and Mark D. Plumbley
- Abstract summary: Convolutional neural networks (CNNs) have shown state-of-the-art performance in various applications.
CNNs are resource-hungry due to their requirement of high computational complexity and memory storage.
Recent efforts toward achieving computational efficiency in CNNs involve filter pruning methods.
- Score: 23.661189257759535
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Convolutional neural networks (CNNs) have shown state-of-the-art performance
in various applications. However, CNNs are resource-hungry due to their
requirement of high computational complexity and memory storage. Recent efforts
toward achieving computational efficiency in CNNs involve filter pruning
methods that eliminate some of the filters in CNNs based on the
\enquote{importance} of the filters. The majority of existing filter pruning
methods are either "active", which use a dataset and generate feature maps to
quantify filter importance, or "passive", which compute filter importance using
entry-wise norm of the filters without involving data. Under a high pruning
ratio where large number of filters are to be pruned from the network, the
entry-wise norm methods eliminate relatively smaller norm filters without
considering the significance of the filters in producing the node output,
resulting in degradation in the performance. To address this, we present a
passive filter pruning method where the filters are pruned based on their
contribution in producing output by considering the operator norm of the
filters. The proposed pruning method generalizes better across various CNNs
compared to that of the entry-wise norm-based pruning methods. In comparison to
the existing active filter pruning methods, the proposed pruning method is at
least 4.5 times faster in computing filter importance and is able to achieve
similar performance compared to that of the active filter pruning methods. The
efficacy of the proposed pruning method is evaluated on audio scene
classification and image classification using various CNNs architecture such as
VGGish, DCASE21_Net, VGG-16 and ResNet-50.
Related papers
- Structured Network Pruning by Measuring Filter-wise Interactions [6.037167142826297]
We propose a structured network pruning approach SNPFI (Structured Network Pruning by measuring Filter-wise Interaction)
During the pruning, the SNPFI can automatically assign the proper sparsity based on the filter utilization strength.
We empirically demonstrate the effectiveness of the SNPFI with several commonly used CNN models.
arXiv Detail & Related papers (2023-07-03T05:26:05Z) - Batch Normalization Tells You Which Filter is Important [49.903610684578716]
We propose a simple yet effective filter pruning method by evaluating the importance of each filter based on the BN parameters of pre-trained CNNs.
The experimental results on CIFAR-10 and ImageNet demonstrate that the proposed method can achieve outstanding performance.
arXiv Detail & Related papers (2021-12-02T12:04:59Z) - Training Compact CNNs for Image Classification using Dynamic-coded
Filter Fusion [139.71852076031962]
We present a novel filter pruning method, dubbed dynamic-coded filter fusion (DCFF)
We derive compact CNNs in a computation-economical and regularization-free manner for efficient image classification.
Our DCFF derives a compact VGGNet-16 with only 72.77M FLOPs and 1.06M parameters while reaching top-1 accuracy of 93.47%.
arXiv Detail & Related papers (2021-07-14T18:07:38Z) - Data Agnostic Filter Gating for Efficient Deep Networks [72.4615632234314]
Current filter pruning methods mainly leverage feature maps to generate important scores for filters and prune those with smaller scores.
In this paper, we propose a data filter pruning method that uses an auxiliary network named Dagger module to induce pruning.
In addition, to help prune filters with certain FLOPs constraints, we leverage an explicit FLOPs-aware regularization to directly promote pruning filters toward target FLOPs.
arXiv Detail & Related papers (2020-10-28T15:26:40Z) - Dependency Aware Filter Pruning [74.69495455411987]
Pruning a proportion of unimportant filters is an efficient way to mitigate the inference cost.
Previous work prunes filters according to their weight norms or the corresponding batch-norm scaling factors.
We propose a novel mechanism to dynamically control the sparsity-inducing regularization so as to achieve the desired sparsity.
arXiv Detail & Related papers (2020-05-06T07:41:22Z) - Filter Grafting for Deep Neural Networks: Reason, Method, and
Cultivation [86.91324735966766]
Filter is the key component in modern convolutional neural networks (CNNs)
In this paper, we introduce filter grafting (textbfMethod) to achieve this goal.
We develop a novel criterion to measure the information of filters and an adaptive weighting strategy to balance the grafted information among networks.
arXiv Detail & Related papers (2020-04-26T08:36:26Z) - Convolutional Neural Network Pruning Using Filter Attenuation [10.282782377635106]
Filters are essential elements in convolutional neural networks (CNNs)
In filter pruning methods, a filter with all of its components, including channels and connections, are removed.
We propose a CNN pruning method based on filter attenuation in which weak filters are not directly removed.
arXiv Detail & Related papers (2020-02-09T06:31:24Z) - Filter Grafting for Deep Neural Networks [71.39169475500324]
Filter grafting aims to improve the representation capability of Deep Neural Networks (DNNs)
We develop an entropy-based criterion to measure the information of filters and an adaptive weighting strategy for balancing the grafted information among networks.
For example, the grafted MobileNetV2 outperforms the non-grafted MobileNetV2 by about 7 percent on CIFAR-100 dataset.
arXiv Detail & Related papers (2020-01-15T03:18:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.