Filter Pruning for Efficient CNNs via Knowledge-driven Differential
Filter Sampler
- URL: http://arxiv.org/abs/2307.00198v1
- Date: Sat, 1 Jul 2023 02:28:41 GMT
- Title: Filter Pruning for Efficient CNNs via Knowledge-driven Differential
Filter Sampler
- Authors: Shaohui Lin, Wenxuan Huang, Jiao Xie, Baochang Zhang, Yunhang Shen,
Zhou Yu, Jungong Han, David Doermann
- Abstract summary: Filter pruning simultaneously accelerates the computation and reduces the memory overhead of CNNs.
We propose a novel Knowledge-driven Differential Filter Sampler(KDFS) with Masked Filter Modeling(MFM) framework for filter pruning.
- Score: 103.97487121678276
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Filter pruning simultaneously accelerates the computation and reduces the
memory overhead of CNNs, which can be effectively applied to edge devices and
cloud services. In this paper, we propose a novel Knowledge-driven Differential
Filter Sampler~(KDFS) with Masked Filter Modeling~(MFM) framework for filter
pruning, which globally prunes the redundant filters based on the prior
knowledge of a pre-trained model in a differential and non-alternative
optimization. Specifically, we design a differential sampler with learnable
sampling parameters to build a binary mask vector for each layer, determining
whether the corresponding filters are redundant. To learn the mask, we
introduce masked filter modeling to construct PCA-like knowledge by aligning
the intermediate features from the pre-trained teacher model and the outputs of
the student decoder taking sampling features as the input. The mask and sampler
are directly optimized by the Gumbel-Softmax Straight-Through Gradient
Estimator in an end-to-end manner in combination with global pruning
constraint, MFM reconstruction error, and dark knowledge. Extensive experiments
demonstrate the proposed KDFS's effectiveness in compressing the base models on
various datasets. For instance, the pruned ResNet-50 on ImageNet achieves
$55.36\%$ computation reduction, and $42.86\%$ parameter reduction, while only
dropping $0.35\%$ Top-1 accuracy, significantly outperforming the
state-of-the-art methods. The code is available at
\url{https://github.com/Osilly/KDFS}.
Related papers
- Building Brain Tumor Segmentation Networks with User-Assisted Filter Estimation and Selection [42.62139206176152]
We present Multi-Step (MS) FLIM, a user-assisted approach to estimating and selecting the most relevant filters from multiple FLIM executions.
MS-FLIM is used only for the first convolutional layer, and the results already indicate improvement over FLIM.
We build a simple U-shaped encoder-decoder network, named sU-Net, for glioblastoma segmentation using T1Gd and FLAIR MRI scans.
arXiv Detail & Related papers (2024-03-19T14:11:26Z) - Asymptotic Soft Cluster Pruning for Deep Neural Networks [5.311178623385279]
Filter pruning method introduces structural sparsity by removing selected filters.
We propose a novel filter pruning method called Asymptotic Soft Cluster Pruning.
Our method can achieve competitive results compared with many state-of-the-art algorithms.
arXiv Detail & Related papers (2022-06-16T13:58:58Z) - Interspace Pruning: Using Adaptive Filter Representations to Improve
Training of Sparse CNNs [69.3939291118954]
Unstructured pruning is well suited to reduce the memory footprint of convolutional neural networks (CNNs)
Standard unstructured pruning (SP) reduces the memory footprint of CNNs by setting filter elements to zero.
We introduce interspace pruning (IP), a general tool to improve existing pruning methods.
arXiv Detail & Related papers (2022-03-15T11:50:45Z) - Batch Normalization Tells You Which Filter is Important [49.903610684578716]
We propose a simple yet effective filter pruning method by evaluating the importance of each filter based on the BN parameters of pre-trained CNNs.
The experimental results on CIFAR-10 and ImageNet demonstrate that the proposed method can achieve outstanding performance.
arXiv Detail & Related papers (2021-12-02T12:04:59Z) - Training Compact CNNs for Image Classification using Dynamic-coded
Filter Fusion [139.71852076031962]
We present a novel filter pruning method, dubbed dynamic-coded filter fusion (DCFF)
We derive compact CNNs in a computation-economical and regularization-free manner for efficient image classification.
Our DCFF derives a compact VGGNet-16 with only 72.77M FLOPs and 1.06M parameters while reaching top-1 accuracy of 93.47%.
arXiv Detail & Related papers (2021-07-14T18:07:38Z) - Unsharp Mask Guided Filtering [53.14430987860308]
The goal of this paper is guided image filtering, which emphasizes the importance of structure transfer during filtering.
We propose a new and simplified formulation of the guided filter inspired by unsharp masking.
Our formulation enjoys a filtering prior to a low-pass filter and enables explicit structure transfer by estimating a single coefficient.
arXiv Detail & Related papers (2021-06-02T19:15:34Z) - Data Agnostic Filter Gating for Efficient Deep Networks [72.4615632234314]
Current filter pruning methods mainly leverage feature maps to generate important scores for filters and prune those with smaller scores.
In this paper, we propose a data filter pruning method that uses an auxiliary network named Dagger module to induce pruning.
In addition, to help prune filters with certain FLOPs constraints, we leverage an explicit FLOPs-aware regularization to directly promote pruning filters toward target FLOPs.
arXiv Detail & Related papers (2020-10-28T15:26:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.