Filter Pruning For CNN With Enhanced Linear Representation Redundancy
- URL: http://arxiv.org/abs/2310.06344v1
- Date: Tue, 10 Oct 2023 06:27:30 GMT
- Title: Filter Pruning For CNN With Enhanced Linear Representation Redundancy
- Authors: Bojue Wang, Chunmei Ma, Bin Liu, Nianbo Liu, Jinqi Zhu
- Abstract summary: We present a data-driven loss function term calculated from the correlation matrix of different feature maps in the same layer, named CCM-loss.
CCM-loss provides us with another universal transcendental mathematical tool besides L*-norm regularization.
In our new strategy, we mainly focus on the consistency and integrality of the information flow in the network.
- Score: 3.853146967741941
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Structured network pruning excels non-structured methods because they can
take advantage of the thriving developed parallel computing techniques. In this
paper, we propose a new structured pruning method. Firstly, to create more
structured redundancy, we present a data-driven loss function term calculated
from the correlation coefficient matrix of different feature maps in the same
layer, named CCM-loss. This loss term can encourage the neural network to learn
stronger linear representation relations between feature maps during the
training from the scratch so that more homogenous parts can be removed later in
pruning. CCM-loss provides us with another universal transcendental
mathematical tool besides L*-norm regularization, which concentrates on
generating zeros, to generate more redundancy but for the different genres.
Furthermore, we design a matching channel selection strategy based on principal
components analysis to exploit the maximum potential ability of CCM-loss. In
our new strategy, we mainly focus on the consistency and integrality of the
information flow in the network. Instead of empirically hard-code the retain
ratio for each layer, our channel selection strategy can dynamically adjust
each layer's retain ratio according to the specific circumstance of a
per-trained model to push the prune ratio to the limit. Notably, on the
Cifar-10 dataset, our method brings 93.64% accuracy for pruned VGG-16 with only
1.40M parameters and 49.60M FLOPs, the pruned ratios for parameters and FLOPs
are 90.6% and 84.2%, respectively. For ResNet-50 trained on the ImageNet
dataset, our approach achieves 42.8% and 47.3% storage and computation
reductions, respectively, with an accuracy of 76.23%. Our code is available at
https://github.com/Bojue-Wang/CCM-LRR.
Related papers
- EGRC-Net: Embedding-induced Graph Refinement Clustering Network [66.44293190793294]
We propose a novel graph clustering network called Embedding-Induced Graph Refinement Clustering Network (EGRC-Net)
EGRC-Net effectively utilizes the learned embedding to adaptively refine the initial graph and enhance the clustering performance.
Our proposed methods consistently outperform several state-of-the-art approaches.
arXiv Detail & Related papers (2022-11-19T09:08:43Z) - A pruning method based on the dissimilarity of angle among channels and
filters [13.878426750493784]
We encode the convolution network to obtain the similarity of different encoding nodes.
We evaluate the connectivity-power among convolutional kernels on the basis of similarity.
We propose Channel Pruning base on the Dissimilarity of Angle (DACP)
arXiv Detail & Related papers (2022-10-29T05:47:57Z) - End-to-End Sensitivity-Based Filter Pruning [49.61707925611295]
We present a sensitivity-based filter pruning algorithm (SbF-Pruner) to learn the importance scores of filters of each layer end-to-end.
Our method learns the scores from the filter weights, enabling it to account for the correlations between the filters of each layer.
arXiv Detail & Related papers (2022-04-15T10:21:05Z) - Basis Scaling and Double Pruning for Efficient Inference in
Network-Based Transfer Learning [1.3467579878240454]
We decompose a convolutional layer into two layers: a convolutional layer with the orthonormal basis vectors as the filters, and a "BasisScalingConv" layer which is responsible for rescaling the features.
We can achieve pruning ratios up to 74.6% for CIFAR-10 and 98.9% for MNIST in model parameters.
arXiv Detail & Related papers (2021-08-06T00:04:02Z) - Group Fisher Pruning for Practical Network Compression [58.25776612812883]
We present a general channel pruning approach that can be applied to various complicated structures.
We derive a unified metric based on Fisher information to evaluate the importance of a single channel and coupled channels.
Our method can be used to prune any structures including those with coupled channels.
arXiv Detail & Related papers (2021-08-02T08:21:44Z) - Model Pruning Based on Quantified Similarity of Feature Maps [5.271060872578571]
We propose a novel theory to find redundant information in three dimensional tensors.
We use this theory to prune convolutional neural networks to enhance the inference speed.
arXiv Detail & Related papers (2021-05-13T02:57:30Z) - Non-Parametric Adaptive Network Pruning [125.4414216272874]
We introduce non-parametric modeling to simplify the algorithm design.
Inspired by the face recognition community, we use a message passing algorithm to obtain an adaptive number of exemplars.
EPruner breaks the dependency on the training data in determining the "important" filters.
arXiv Detail & Related papers (2021-01-20T06:18:38Z) - ACP: Automatic Channel Pruning via Clustering and Swarm Intelligence
Optimization for CNN [6.662639002101124]
convolutional neural network (CNN) gets deeper and wider in recent years.
Existing magnitude-based pruning methods are efficient, but the performance of the compressed network is unpredictable.
We propose a novel automatic channel pruning method (ACP)
ACP is evaluated against several state-of-the-art CNNs on three different classification datasets.
arXiv Detail & Related papers (2021-01-16T08:56:38Z) - Layer Pruning via Fusible Residual Convolutional Block for Deep Neural
Networks [15.64167076052513]
layer pruning has less inference time and runtime memory usage when the same FLOPs and number of parameters are pruned.
We propose a simple layer pruning method using residual convolutional block (ResConv)
Our pruning method achieves excellent performance of compression and acceleration over the state-thearts on different datasets.
arXiv Detail & Related papers (2020-11-29T12:51:16Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z) - Filter Sketch for Network Pruning [184.41079868885265]
We propose a novel network pruning approach by information preserving of pre-trained network weights (filters)
Our approach, referred to as FilterSketch, encodes the second-order information of pre-trained weights.
Experiments on CIFAR-10 show that FilterSketch reduces 63.3% of FLOPs and prunes 59.9% of network parameters with negligible accuracy cost.
arXiv Detail & Related papers (2020-01-23T13:57:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.