Pruning Filter in Filter
- URL: http://arxiv.org/abs/2009.14410v3
- Date: Wed, 9 Dec 2020 08:35:21 GMT
- Title: Pruning Filter in Filter
- Authors: Fanxu Meng, Hao Cheng, Ke Li, Huixiang Luo, Xiaowei Guo, Guangming Lu,
Xing Sun
- Abstract summary: Pruning has become a very powerful and effective technique to compress and accelerate modern neural networks.
We propose to prune the filter in the filter to achieve finer granularity than traditional filter pruning methods.
We demonstrate that SWP is more effective compared to the previous FP-based methods and achieves the state-of-art pruning ratio on CIFAR-10 and ImageNet datasets.
- Score: 38.6403556260338
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pruning has become a very powerful and effective technique to compress and
accelerate modern neural networks. Existing pruning methods can be grouped into
two categories: filter pruning (FP) and weight pruning (WP). FP wins at
hardware compatibility but loses at the compression ratio compared with WP. To
converge the strength of both methods, we propose to prune the filter in the
filter. Specifically, we treat a filter $F \in \mathbb{R}^{C\times K\times K}$
as $K \times K$ stripes, i.e., $1\times 1$ filters $\in \mathbb{R}^{C}$, then
by pruning the stripes instead of the whole filter, we can achieve finer
granularity than traditional FP while being hardware friendly. We term our
method as SWP (\emph{Stripe-Wise Pruning}). SWP is implemented by introducing a
novel learnable matrix called Filter Skeleton, whose values reflect the shape
of each filter. As some recent work has shown that the pruned architecture is
more crucial than the inherited important weights, we argue that the
architecture of a single filter, i.e., the shape, also matters. Through
extensive experiments, we demonstrate that SWP is more effective compared to
the previous FP-based methods and achieves the state-of-art pruning ratio on
CIFAR-10 and ImageNet datasets without obvious accuracy drop. Code is available
at https://github.com/fxmeng/Pruning-Filter-in-Filter
Related papers
- Filter Pruning for Efficient CNNs via Knowledge-driven Differential
Filter Sampler [103.97487121678276]
Filter pruning simultaneously accelerates the computation and reduces the memory overhead of CNNs.
We propose a novel Knowledge-driven Differential Filter Sampler(KDFS) with Masked Filter Modeling(MFM) framework for filter pruning.
arXiv Detail & Related papers (2023-07-01T02:28:41Z) - Pruning Networks with Cross-Layer Ranking & k-Reciprocal Nearest Filters [151.2423480789271]
A novel pruning method, termed CLR-RNF, is proposed for filter-level network pruning.
We conduct image classification on CIFAR-10 and ImageNet to demonstrate the superiority of our CLR-RNF over the state-of-the-arts.
arXiv Detail & Related papers (2022-02-15T04:53:24Z) - One-Sided Box Filter for Edge Preserving Image Smoothing [9.67565473617028]
We present a one-sided box filter that can smooth the signal but keep the discontinuous features in the signal.
More specifically, we perform box filter on eight one-sided windows, leading to a one-sided box filter that can preserve corners and edges.
arXiv Detail & Related papers (2021-08-11T04:22:38Z) - Unsharp Mask Guided Filtering [53.14430987860308]
The goal of this paper is guided image filtering, which emphasizes the importance of structure transfer during filtering.
We propose a new and simplified formulation of the guided filter inspired by unsharp masking.
Our formulation enjoys a filtering prior to a low-pass filter and enables explicit structure transfer by estimating a single coefficient.
arXiv Detail & Related papers (2021-06-02T19:15:34Z) - Deep Model Compression based on the Training History [13.916984628784768]
We propose a novel History Based Filter Pruning (HBFP) method that utilizes network training history for filter pruning.
The proposed pruning method outperforms the state-of-the-art in terms of FLOPs reduction (floating-point operations) by 97.98%, 83.42%, 78.43%, and 74.95% for LeNet-5, VGG-16, ResNet-56, and ResNet-110 models, respectively.
arXiv Detail & Related papers (2021-01-30T06:04:21Z) - Data Agnostic Filter Gating for Efficient Deep Networks [72.4615632234314]
Current filter pruning methods mainly leverage feature maps to generate important scores for filters and prune those with smaller scores.
In this paper, we propose a data filter pruning method that uses an auxiliary network named Dagger module to induce pruning.
In addition, to help prune filters with certain FLOPs constraints, we leverage an explicit FLOPs-aware regularization to directly promote pruning filters toward target FLOPs.
arXiv Detail & Related papers (2020-10-28T15:26:40Z) - Filter Grafting for Deep Neural Networks: Reason, Method, and
Cultivation [86.91324735966766]
Filter is the key component in modern convolutional neural networks (CNNs)
In this paper, we introduce filter grafting (textbfMethod) to achieve this goal.
We develop a novel criterion to measure the information of filters and an adaptive weighting strategy to balance the grafted information among networks.
arXiv Detail & Related papers (2020-04-26T08:36:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.