Can Unstructured Pruning Reduce the Depth in Deep Neural Networks?
- URL: http://arxiv.org/abs/2308.06619v2
- Date: Fri, 18 Aug 2023 13:00:23 GMT
- Title: Can Unstructured Pruning Reduce the Depth in Deep Neural Networks?
- Authors: Zhu Liao, Victor Qu\'etu, Van-Tam Nguyen, Enzo Tartaglione
- Abstract summary: Pruning is a widely used technique for reducing the size of deep neural networks while maintaining their performance.
In this study, we introduce EGP, an innovative Entropy Guided Pruning algorithm aimed at reducing the size of deep neural networks while preserving their performance.
- Score: 5.869633234882029
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pruning is a widely used technique for reducing the size of deep neural
networks while maintaining their performance. However, such a technique,
despite being able to massively compress deep models, is hardly able to remove
entire layers from a model (even when structured): is this an addressable task?
In this study, we introduce EGP, an innovative Entropy Guided Pruning algorithm
aimed at reducing the size of deep neural networks while preserving their
performance. The key focus of EGP is to prioritize pruning connections in
layers with low entropy, ultimately leading to their complete removal. Through
extensive experiments conducted on popular models like ResNet-18 and Swin-T,
our findings demonstrate that EGP effectively compresses deep neural networks
while maintaining competitive performance levels. Our results not only shed
light on the underlying mechanism behind the advantages of unstructured
pruning, but also pave the way for further investigations into the intricate
relationship between entropy, pruning techniques, and deep learning
performance. The EGP algorithm and its insights hold great promise for
advancing the field of network compression and optimization. The source code
for EGP is released open-source.
Related papers
- NEPENTHE: Entropy-Based Pruning as a Neural Network Depth's Reducer [5.373015313199385]
We propose an eNtropy-basEd Pruning as a nEural Network depTH's rEducer to alleviate deep neural networks' computational burden.
We validate our approach on popular architectures such as MobileNet and Swin-T.
arXiv Detail & Related papers (2024-04-24T09:12:04Z) - Regularized PolyKervNets: Optimizing Expressiveness and Efficiency for
Private Inference in Deep Neural Networks [0.0]
We focus on PolyKervNets, a technique known for offering improved dynamic approximations in smaller networks.
Our primary objective is to empirically explore optimization-based training recipes to enhance the performance of PolyKervNets in larger networks.
arXiv Detail & Related papers (2023-12-23T11:37:18Z) - Structured Bayesian Compression for Deep Neural Networks Based on The
Turbo-VBI Approach [23.729955669774977]
In most existing pruning methods, surviving neurons are randomly connected in the neural network without any structure.
We propose a three-layer hierarchical prior to promote a more regular sparse structure during pruning.
We derive an efficient Turbo-variational Bayesian inferencing (Turbo-VBI) algorithm to solve the resulting model compression problem.
arXiv Detail & Related papers (2023-02-21T07:12:36Z) - SAR Despeckling Using Overcomplete Convolutional Networks [53.99620005035804]
despeckling is an important problem in remote sensing as speckle degrades SAR images.
Recent studies show that convolutional neural networks(CNNs) outperform classical despeckling methods.
This study employs an overcomplete CNN architecture to focus on learning low-level features by restricting the receptive field.
We show that the proposed network improves despeckling performance compared to recent despeckling methods on synthetic and real SAR images.
arXiv Detail & Related papers (2022-05-31T15:55:37Z) - Analytically Tractable Inference in Deep Neural Networks [0.0]
Tractable Approximate Inference (TAGI) algorithm was shown to be a viable and scalable alternative to backpropagation for shallow fully-connected neural networks.
We are demonstrating how TAGI matches or exceeds the performance of backpropagation, for training classic deep neural network architectures.
arXiv Detail & Related papers (2021-03-09T14:51:34Z) - Learning N:M Fine-grained Structured Sparse Neural Networks From Scratch [75.69506249886622]
Sparsity in Deep Neural Networks (DNNs) has been widely studied to compress and accelerate the models on resource-constrained environments.
In this paper, we are the first to study training from scratch an N:M fine-grained structured sparse network.
arXiv Detail & Related papers (2021-02-08T05:55:47Z) - Finite Versus Infinite Neural Networks: an Empirical Study [69.07049353209463]
kernel methods outperform fully-connected finite-width networks.
Centered and ensembled finite networks have reduced posterior variance.
Weight decay and the use of a large learning rate break the correspondence between finite and infinite networks.
arXiv Detail & Related papers (2020-07-31T01:57:47Z) - ALF: Autoencoder-based Low-rank Filter-sharing for Efficient
Convolutional Neural Networks [63.91384986073851]
We propose the autoencoder-based low-rank filter-sharing technique technique (ALF)
ALF shows a reduction of 70% in network parameters, 61% in operations and 41% in execution time, with minimal loss in accuracy.
arXiv Detail & Related papers (2020-07-27T09:01:22Z) - Rectified Linear Postsynaptic Potential Function for Backpropagation in
Deep Spiking Neural Networks [55.0627904986664]
Spiking Neural Networks (SNNs) usetemporal spike patterns to represent and transmit information, which is not only biologically realistic but also suitable for ultra-low-power event-driven neuromorphic implementation.
This paper investigates the contribution of spike timing dynamics to information encoding, synaptic plasticity and decision making, providing a new perspective to design of future DeepSNNs and neuromorphic hardware systems.
arXiv Detail & Related papers (2020-03-26T11:13:07Z) - Robust Pruning at Initialization [61.30574156442608]
A growing need for smaller, energy-efficient, neural networks to be able to use machine learning applications on devices with limited computational resources.
For Deep NNs, such procedures remain unsatisfactory as the resulting pruned networks can be difficult to train and, for instance, they do not prevent one layer from being fully pruned.
arXiv Detail & Related papers (2020-02-19T17:09:50Z) - Lossless Compression of Deep Neural Networks [17.753357839478575]
Deep neural networks have been successful in many predictive modeling tasks, such as image and language recognition.
It is challenging to deploy these networks under limited computational resources, such as in mobile devices.
We introduce an algorithm that removes units and layers of a neural network while not changing the output that is produced.
arXiv Detail & Related papers (2020-01-01T15:04:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.