L2PF -- Learning to Prune Faster
- URL: http://arxiv.org/abs/2101.02663v1
- Date: Thu, 7 Jan 2021 18:13:37 GMT
- Title: L2PF -- Learning to Prune Faster
- Authors: Manoj-Rohit Vemparala, Nael Fasfous, Alexander Frickenstein, Mhd Ali
Moraly, Aquib Jamal, Lukas Frickenstein, Christian Unger, Naveen-Shankar
Nagaraja, Walter Stechele
- Abstract summary: We present a multi-task, try-and-learn method, discretely learning redundant filters of the CNN and a continuous action of how long the layers have to be fine-tuned.
For ResNet20, we have achieved a compression ratio of 3.84 x with minimal accuracy degradation.
Compared to the state-of-the-art pruning method, we reduced the GPU hours by 1.71 x.
- Score: 57.32153461504626
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Various applications in the field of autonomous driving are based on
convolutional neural networks (CNNs), especially for processing camera data.
The optimization of such CNNs is a major challenge in continuous development.
Newly learned features must be brought into vehicles as quickly as possible,
and as such, it is not feasible to spend redundant GPU hours during
compression. In this context, we present Learning to Prune Faster which details
a multi-task, try-and-learn method, discretely learning redundant filters of
the CNN and a continuous action of how long the layers have to be fine-tuned.
This allows us to significantly speed up the convergence process of learning
how to find an embedded-friendly filter-wise pruned CNN. For ResNet20, we have
achieved a compression ratio of 3.84 x with minimal accuracy degradation.
Compared to the state-of-the-art pruning method, we reduced the GPU hours by
1.71 x.
Related papers
- SparseTem: Boosting the Efficiency of CNN-Based Video Encoders by Exploiting Temporal Continuity [15.872209884833977]
We propose a memory-efficient scheduling method to eliminate memory overhead and an online adjustment mechanism to minimize accuracy degradation.
SparseTem achieves speedup of 1.79x for EfficientDet and 4.72x for CRNN, with minimal accuracy drop and no additional memory overhead.
arXiv Detail & Related papers (2024-10-28T07:13:25Z) - Transferability of Convolutional Neural Networks in Stationary Learning
Tasks [96.00428692404354]
We introduce a novel framework for efficient training of convolutional neural networks (CNNs) for large-scale spatial problems.
We show that a CNN trained on small windows of such signals achieves a nearly performance on much larger windows without retraining.
Our results show that the CNN is able to tackle problems with many hundreds of agents after being trained with fewer than ten.
arXiv Detail & Related papers (2023-07-21T13:51:45Z) - Self Similarity Matrix based CNN Filter Pruning [1.6799377888527687]
We tackle the problem of CNN model pruning with the help of Self-Similarity Matrix (SSM) computed from the 2D CNN filters.
We propose two novel algorithms to rank and prune redundant filters which contribute similar activation maps to the output.
arXiv Detail & Related papers (2022-11-03T13:47:44Z) - Compressing CNN Kernels for Videos Using Tucker Decompositions: Towards
Lightweight CNN Applications [2.191505742658975]
Convolutional Neural Networks (CNN) are the state-of-theart in the field of visual computing.
A major problem with CNNs is the large number of floating point operations (FLOPs) required to perform convolutions for large inputs.
We propose a Tuckerdecomposition to compress the convolutional kernel of a pre-trained network for images.
arXiv Detail & Related papers (2022-03-10T11:53:53Z) - An Acceleration Method Based on Deep Learning and Multilinear Feature
Space [0.0]
This paper presents an alternative approach based on the Multilinear Feature Space (MFS) method resorting to transfer learning from large CNN architectures.
The proposed method uses CNNs to generate feature maps, although it does not work as complexity reduction approach.
Our method, named AMFC, uses the transfer learning from pre-trained CNN to reduce the classification time of new sample image, with minimal accuracy loss.
arXiv Detail & Related papers (2021-10-16T23:49:12Z) - Content-Aware Convolutional Neural Networks [98.97634685964819]
Convolutional Neural Networks (CNNs) have achieved great success due to the powerful feature learning ability of convolution layers.
We propose a Content-aware Convolution (CAC) that automatically detects the smooth windows and applies a 1x1 convolutional kernel to replace the original large kernel.
arXiv Detail & Related papers (2021-06-30T03:54:35Z) - 3D CNNs with Adaptive Temporal Feature Resolutions [83.43776851586351]
Similarity Guided Sampling (SGS) module can be plugged into any existing 3D CNN architecture.
SGS empowers 3D CNNs by learning the similarity of temporal features and grouping similar features together.
Our evaluations show that the proposed module improves the state-of-the-art by reducing the computational cost (GFLOPs) by half while preserving or even improving the accuracy.
arXiv Detail & Related papers (2020-11-17T14:34:05Z) - When deep learning models on GPU can be accelerated by taking advantage
of unstructured sparsity [0.0]
This paper is focused on the improvement the efficiency of the sparse convolutional neural networks (CNNs) layers on graphic processing units ( GPU)
The modern CNN models need megabytes of coefficients and needed millions MAC operations to perform convolution.
We show when is worth using a direct sparse operation to speed-up the calculation of the convolution layers.
arXiv Detail & Related papers (2020-11-12T10:13:48Z) - RT3D: Achieving Real-Time Execution of 3D Convolutional Neural Networks
on Mobile Devices [57.877112704841366]
This paper proposes RT3D, a model compression and mobile acceleration framework for 3D CNNs.
For the first time, real-time execution of 3D CNNs is achieved on off-the-shelf mobiles.
arXiv Detail & Related papers (2020-07-20T02:05:32Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.