Filter Sketch for Network Pruning
- URL: http://arxiv.org/abs/2001.08514v4
- Date: Tue, 25 May 2021 03:12:53 GMT
- Title: Filter Sketch for Network Pruning
- Authors: Mingbao Lin, Liujuan Cao, Shaojie Li, Qixiang Ye, Yonghong Tian,
Jianzhuang Liu, Qi Tian, Rongrong Ji
- Abstract summary: We propose a novel network pruning approach by information preserving of pre-trained network weights (filters)
Our approach, referred to as FilterSketch, encodes the second-order information of pre-trained weights.
Experiments on CIFAR-10 show that FilterSketch reduces 63.3% of FLOPs and prunes 59.9% of network parameters with negligible accuracy cost.
- Score: 184.41079868885265
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel network pruning approach by information preserving of
pre-trained network weights (filters). Network pruning with the information
preserving is formulated as a matrix sketch problem, which is efficiently
solved by the off-the-shelf Frequent Direction method. Our approach, referred
to as FilterSketch, encodes the second-order information of pre-trained
weights, which enables the representation capacity of pruned networks to be
recovered with a simple fine-tuning procedure. FilterSketch requires neither
training from scratch nor data-driven iterative optimization, leading to a
several-orders-of-magnitude reduction of time cost in the optimization of
pruning. Experiments on CIFAR-10 show that FilterSketch reduces 63.3% of FLOPs
and prunes 59.9% of network parameters with negligible accuracy cost for
ResNet-110. On ILSVRC-2012, it reduces 45.5% of FLOPs and removes 43.0% of
parameters with only 0.69% accuracy drop for ResNet-50. Our code and pruned
models can be found at https://github.com/lmbxmu/FilterSketch.
Related papers
- Filter Pruning for Efficient CNNs via Knowledge-driven Differential
Filter Sampler [103.97487121678276]
Filter pruning simultaneously accelerates the computation and reduces the memory overhead of CNNs.
We propose a novel Knowledge-driven Differential Filter Sampler(KDFS) with Masked Filter Modeling(MFM) framework for filter pruning.
arXiv Detail & Related papers (2023-07-01T02:28:41Z) - Pruning by Active Attention Manipulation [49.61707925611295]
Filter pruning of a CNN is typically achieved by applying discrete masks on the CNN's filter weights or activation maps, post-training.
Here, we present a new filter-importance-scoring concept named pruning by active attention manipulation (PAAM)
PAAM learns analog filter scores from the filter weights by optimizing a cost function regularized by an additive term in the scores.
arXiv Detail & Related papers (2022-10-20T09:17:02Z) - End-to-End Sensitivity-Based Filter Pruning [49.61707925611295]
We present a sensitivity-based filter pruning algorithm (SbF-Pruner) to learn the importance scores of filters of each layer end-to-end.
Our method learns the scores from the filter weights, enabling it to account for the correlations between the filters of each layer.
arXiv Detail & Related papers (2022-04-15T10:21:05Z) - Pruning Networks with Cross-Layer Ranking & k-Reciprocal Nearest Filters [151.2423480789271]
A novel pruning method, termed CLR-RNF, is proposed for filter-level network pruning.
We conduct image classification on CIFAR-10 and ImageNet to demonstrate the superiority of our CLR-RNF over the state-of-the-arts.
arXiv Detail & Related papers (2022-02-15T04:53:24Z) - Pruning-aware Sparse Regularization for Network Pruning [38.13617051756663]
Structural neural network pruning aims to remove the redundant channels in the deep convolutional neural networks (CNNs)
In this paper, we analyze these sparsity-training-based methods and find that the regularization of unpruned channels is unnecessary.
We propose a novel pruning method, named MaskSparsity, with pruning-aware sparse regularization.
arXiv Detail & Related papers (2022-01-18T07:19:23Z) - Network Compression via Central Filter [9.585818883354449]
We propose a novel filter pruning method, Central Filter (CF), which suggests a filter is approximately equal to a set of other filters after appropriate adjustments.
CF yields state-of-the-art performance on various benchmark networks and datasets.
arXiv Detail & Related papers (2021-12-10T12:51:04Z) - Automatic Neural Network Pruning that Efficiently Preserves the Model
Accuracy [2.538209532048867]
pruning filters is a common solution, but most existing pruning methods do not preserve the model accuracy efficiently.
We propose an automatic pruning method that learns which neurons to preserve in order to maintain the model accuracy while reducing the FLOPs to a predefined target.
We achieve a 52.00% FLOPs reduction on ResNet-50, with a Top-1 accuracy of 47.51% after pruning and a state-of-the-art (SOTA) accuracy of 76.63% after finetuning.
arXiv Detail & Related papers (2021-11-18T11:29:35Z) - Deep Model Compression based on the Training History [13.916984628784768]
We propose a novel History Based Filter Pruning (HBFP) method that utilizes network training history for filter pruning.
The proposed pruning method outperforms the state-of-the-art in terms of FLOPs reduction (floating-point operations) by 97.98%, 83.42%, 78.43%, and 74.95% for LeNet-5, VGG-16, ResNet-56, and ResNet-110 models, respectively.
arXiv Detail & Related papers (2021-01-30T06:04:21Z) - Non-Parametric Adaptive Network Pruning [125.4414216272874]
We introduce non-parametric modeling to simplify the algorithm design.
Inspired by the face recognition community, we use a message passing algorithm to obtain an adaptive number of exemplars.
EPruner breaks the dependency on the training data in determining the "important" filters.
arXiv Detail & Related papers (2021-01-20T06:18:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.