Self Similarity Matrix based CNN Filter Pruning
- URL: http://arxiv.org/abs/2211.01814v1
- Date: Thu, 3 Nov 2022 13:47:44 GMT
- Title: Self Similarity Matrix based CNN Filter Pruning
- Authors: S Rakshith, Jayesh Rajkumar Vachhani, Sourabh Vasant Gothe, and
Rishabh Khurana
- Abstract summary: We tackle the problem of CNN model pruning with the help of Self-Similarity Matrix (SSM) computed from the 2D CNN filters.
We propose two novel algorithms to rank and prune redundant filters which contribute similar activation maps to the output.
- Score: 1.6799377888527687
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, most of the deep learning solutions are targeted to be
deployed in mobile devices. This makes the need for development of lightweight
models all the more imminent. Another solution is to optimize and prune regular
deep learning models. In this paper, we tackle the problem of CNN model pruning
with the help of Self-Similarity Matrix (SSM) computed from the 2D CNN filters.
We propose two novel algorithms to rank and prune redundant filters which
contribute similar activation maps to the output. One of the key features of
our method is that there is no need of finetuning after training the model.
Both the training and pruning process is completed simultaneously. We benchmark
our method on two of the most popular CNN models - ResNet and VGG and record
their performance on the CIFAR-10 dataset.
Related papers
- Truncated Consistency Models [57.50243901368328]
Training consistency models requires learning to map all intermediate points along PF ODE trajectories to their corresponding endpoints.
We empirically find that this training paradigm limits the one-step generation performance of consistency models.
We propose a new parameterization of the consistency function and a two-stage training procedure that prevents the truncated-time training from collapsing to a trivial solution.
arXiv Detail & Related papers (2024-10-18T22:38:08Z) - Reusing Convolutional Neural Network Models through Modularization and
Composition [22.823870645316397]
We propose two modularization approaches named CNNSplitter and GradSplitter.
CNNSplitter decomposes a trained convolutional neural network (CNN) model into $N$ small reusable modules.
The resulting modules can be reused to patch existing CNN models or build new CNN models through composition.
arXiv Detail & Related papers (2023-11-08T03:18:49Z) - A Lightweight CNN-Transformer Model for Learning Traveling Salesman
Problems [0.0]
CNN-Transformer model is able to better learn spatial features from input data using a CNN embedding layer.
The proposed model exhibits the best performance in real-world datasets.
arXiv Detail & Related papers (2023-05-03T04:28:10Z) - An Acceleration Method Based on Deep Learning and Multilinear Feature
Space [0.0]
This paper presents an alternative approach based on the Multilinear Feature Space (MFS) method resorting to transfer learning from large CNN architectures.
The proposed method uses CNNs to generate feature maps, although it does not work as complexity reduction approach.
Our method, named AMFC, uses the transfer learning from pre-trained CNN to reduce the classification time of new sample image, with minimal accuracy loss.
arXiv Detail & Related papers (2021-10-16T23:49:12Z) - Learning from Images: Proactive Caching with Parallel Convolutional
Neural Networks [94.85780721466816]
A novel framework for proactive caching is proposed in this paper.
It combines model-based optimization with data-driven techniques by transforming an optimization problem into a grayscale image.
Numerical results show that the proposed scheme can reduce 71.6% computation time with only 0.8% additional performance cost.
arXiv Detail & Related papers (2021-08-15T21:32:47Z) - Image Restoration by Deep Projected GSURE [115.57142046076164]
Ill-posed inverse problems appear in many image processing applications, such as deblurring and super-resolution.
We propose a new image restoration framework that is based on minimizing a loss function that includes a "projected-version" of the Generalized SteinUnbiased Risk Estimator (GSURE) and parameterization of the latent image by a CNN.
arXiv Detail & Related papers (2021-02-04T08:52:46Z) - L2PF -- Learning to Prune Faster [57.32153461504626]
We present a multi-task, try-and-learn method, discretely learning redundant filters of the CNN and a continuous action of how long the layers have to be fine-tuned.
For ResNet20, we have achieved a compression ratio of 3.84 x with minimal accuracy degradation.
Compared to the state-of-the-art pruning method, we reduced the GPU hours by 1.71 x.
arXiv Detail & Related papers (2021-01-07T18:13:37Z) - When deep learning models on GPU can be accelerated by taking advantage
of unstructured sparsity [0.0]
This paper is focused on the improvement the efficiency of the sparse convolutional neural networks (CNNs) layers on graphic processing units ( GPU)
The modern CNN models need megabytes of coefficients and needed millions MAC operations to perform convolution.
We show when is worth using a direct sparse operation to speed-up the calculation of the convolution layers.
arXiv Detail & Related papers (2020-11-12T10:13:48Z) - Approximated Bilinear Modules for Temporal Modeling [116.6506871576514]
Two-layers in CNNs can be converted to temporal bilinear modules by adding an auxiliary-branch sampling.
Our models can outperform most state-of-the-art methods on SomethingSomething v1 and v2 datasets without pretraining.
arXiv Detail & Related papers (2020-07-25T09:07:35Z) - Stepwise Model Selection for Sequence Prediction via Deep Kernel
Learning [100.83444258562263]
We propose a novel Bayesian optimization (BO) algorithm to tackle the challenge of model selection in this setting.
In order to solve the resulting multiple black-box function optimization problem jointly and efficiently, we exploit potential correlations among black-box functions.
We are the first to formulate the problem of stepwise model selection (SMS) for sequence prediction, and to design and demonstrate an efficient joint-learning algorithm for this purpose.
arXiv Detail & Related papers (2020-01-12T09:42:19Z) - CNNTOP: a CNN-based Trajectory Owner Prediction Method [1.3793594968500604]
Trajectory owner prediction is the basis for many applications such as personalized recommendation, urban planning.
Existing methods mainly employ RNNs to model trajectories semantically.
We propose a CNN-based Trajectory Owner Prediction (CNNTOP) method.
arXiv Detail & Related papers (2020-01-05T07:58:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.