Improve Convolutional Neural Network Pruning by Maximizing Filter
Variety
- URL: http://arxiv.org/abs/2203.05807v1
- Date: Fri, 11 Mar 2022 09:00:59 GMT
- Title: Improve Convolutional Neural Network Pruning by Maximizing Filter
Variety
- Authors: Nathan Hubens, Matei Mancas, Bernard Gosselin, Marius Preda, Titus
Zaharia
- Abstract summary: Neural network pruning is a widely used strategy for reducing model storage and computing requirements.
Common pruning criteria, such as l1-norm or movement, usually do not consider the individual utility of filters.
We present a technique solving those two issues, and which can be appended to any pruning criteria.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural network pruning is a widely used strategy for reducing model storage
and computing requirements. It allows to lower the complexity of the network by
introducing sparsity in the weights. Because taking advantage of sparse
matrices is still challenging, pruning is often performed in a structured way,
i.e. removing entire convolution filters in the case of ConvNets, according to
a chosen pruning criteria. Common pruning criteria, such as l1-norm or
movement, usually do not consider the individual utility of filters, which may
lead to: (1) the removal of filters exhibiting rare, thus important and
discriminative behaviour, and (2) the retaining of filters with redundant
information. In this paper, we present a technique solving those two issues,
and which can be appended to any pruning criteria. This technique ensures that
the criteria of selection focuses on redundant filters, while retaining the
rare ones, thus maximizing the variety of remaining filters. The experimental
results, carried out on different datasets (CIFAR-10, CIFAR-100 and
CALTECH-101) and using different architectures (VGG-16 and ResNet-18)
demonstrate that it is possible to achieve similar sparsity levels while
maintaining a higher performance when appending our filter selection technique
to pruning criteria. Moreover, we assess the quality of the found sparse
sub-networks by applying the Lottery Ticket Hypothesis and find that the
addition of our method allows to discover better performing tickets in most
cases
Related papers
- A Greedy Hierarchical Approach to Whole-Network Filter-Pruning in CNNs [2.188091591747149]
Whole-network filter pruning algorithms prune varying fractions of filters from each layer, hence providing greater flexibility.
This paper proposes a two-level hierarchical approach for whole-network filter pruning which is efficient and uses the classification loss as the final criterion.
Our method reduces the RAM requirement for ResNext101 from 7.6 GB to 1.5 GB and achieves a 94% reduction in FLOPS without losing accuracy on CIFAR-10.
arXiv Detail & Related papers (2024-08-22T03:59:57Z) - Efficient CNNs via Passive Filter Pruning [23.661189257759535]
Convolutional neural networks (CNNs) have shown state-of-the-art performance in various applications.
CNNs are resource-hungry due to their requirement of high computational complexity and memory storage.
Recent efforts toward achieving computational efficiency in CNNs involve filter pruning methods.
arXiv Detail & Related papers (2023-04-05T09:19:19Z) - Asymptotic Soft Cluster Pruning for Deep Neural Networks [5.311178623385279]
Filter pruning method introduces structural sparsity by removing selected filters.
We propose a novel filter pruning method called Asymptotic Soft Cluster Pruning.
Our method can achieve competitive results compared with many state-of-the-art algorithms.
arXiv Detail & Related papers (2022-06-16T13:58:58Z) - Pruning Networks with Cross-Layer Ranking & k-Reciprocal Nearest Filters [151.2423480789271]
A novel pruning method, termed CLR-RNF, is proposed for filter-level network pruning.
We conduct image classification on CIFAR-10 and ImageNet to demonstrate the superiority of our CLR-RNF over the state-of-the-arts.
arXiv Detail & Related papers (2022-02-15T04:53:24Z) - Learning Versatile Convolution Filters for Efficient Visual Recognition [125.34595948003745]
This paper introduces versatile filters to construct efficient convolutional neural networks.
We conduct theoretical analysis on network complexity and an efficient convolution scheme is introduced.
Experimental results on benchmark datasets and neural networks demonstrate that our versatile filters are able to achieve comparable accuracy as that of original filters.
arXiv Detail & Related papers (2021-09-20T06:07:14Z) - Data Agnostic Filter Gating for Efficient Deep Networks [72.4615632234314]
Current filter pruning methods mainly leverage feature maps to generate important scores for filters and prune those with smaller scores.
In this paper, we propose a data filter pruning method that uses an auxiliary network named Dagger module to induce pruning.
In addition, to help prune filters with certain FLOPs constraints, we leverage an explicit FLOPs-aware regularization to directly promote pruning filters toward target FLOPs.
arXiv Detail & Related papers (2020-10-28T15:26:40Z) - Dependency Aware Filter Pruning [74.69495455411987]
Pruning a proportion of unimportant filters is an efficient way to mitigate the inference cost.
Previous work prunes filters according to their weight norms or the corresponding batch-norm scaling factors.
We propose a novel mechanism to dynamically control the sparsity-inducing regularization so as to achieve the desired sparsity.
arXiv Detail & Related papers (2020-05-06T07:41:22Z) - Convolution-Weight-Distribution Assumption: Rethinking the Criteria of
Channel Pruning [90.2947802490534]
We find two blind spots in the study of pruning criteria.
The ranks of filters'Importance Score are almost identical, resulting in similar pruned structures.
The filters'Importance Score measured by some pruning criteria are too close to distinguish the network redundancy well.
arXiv Detail & Related papers (2020-04-24T09:54:21Z) - Group Sparsity: The Hinge Between Filter Pruning and Decomposition for
Network Compression [145.04742985050808]
We analyze two popular network compression techniques, i.e. filter pruning and low-rank decomposition, in a unified sense.
By changing the way the sparsity regularization is enforced, filter pruning and low-rank decomposition can be derived accordingly.
Our approach proves its potential as it compares favorably to the state-of-the-art on several benchmarks.
arXiv Detail & Related papers (2020-03-19T17:57:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.