A Framework For Pruning Deep Neural Networks Using Energy-Based Models
- URL: http://arxiv.org/abs/2102.13188v1
- Date: Thu, 25 Feb 2021 21:44:19 GMT
- Title: A Framework For Pruning Deep Neural Networks Using Energy-Based Models
- Authors: Hojjat Salehinejad, Shahrokh Valaee
- Abstract summary: A typical deep neural network (DNN) has a large number of trainable parameters.
We propose a framework for pruning DNNs based on a population-based global optimization method.
Experiments on ResNets, AlexNet, and SqueezeNet show a pruning rate of more than $50%$ of the trainable parameters.
- Score: 45.4796383952516
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A typical deep neural network (DNN) has a large number of trainable
parameters. Choosing a network with proper capacity is challenging and
generally a larger network with excessive capacity is trained. Pruning is an
established approach to reducing the number of parameters in a DNN. In this
paper, we propose a framework for pruning DNNs based on a population-based
global optimization method. This framework can use any pruning objective
function. As a case study, we propose a simple but efficient objective function
based on the concept of energy-based models. Our experiments on ResNets,
AlexNet, and SqueezeNet for the CIFAR-10 and CIFAR-100 datasets show a pruning
rate of more than $50\%$ of the trainable parameters with approximately $<5\%$
and $<1\%$ drop of Top-1 and Top-5 classification accuracy, respectively.
Related papers
- Just How Flexible are Neural Networks in Practice? [89.80474583606242]
It is widely believed that a neural network can fit a training set containing at least as many samples as it has parameters.
In practice, however, we only find solutions via our training procedure, including the gradient and regularizers, limiting flexibility.
arXiv Detail & Related papers (2024-06-17T12:24:45Z) - Learning Activation Functions for Sparse Neural Networks [12.234742322758418]
Sparse Neural Networks (SNNs) can potentially demonstrate similar performance to their dense counterparts.
However, the accuracy drop incurred by SNNs, especially at high pruning ratios, can be an issue in critical deployment conditions.
We focus on learning a novel way to tune activation functions for sparse networks.
arXiv Detail & Related papers (2023-05-18T13:30:29Z) - Rewarded meta-pruning: Meta Learning with Rewards for Channel Pruning [19.978542231976636]
This paper proposes a novel method to reduce the parameters and FLOPs for computational efficiency in deep learning models.
We introduce accuracy and efficiency coefficients to control the trade-off between the accuracy of the network and its computing efficiency.
arXiv Detail & Related papers (2023-01-26T12:32:01Z) - Neural Capacitance: A New Perspective of Neural Network Selection via
Edge Dynamics [85.31710759801705]
Current practice requires expensive computational costs in model training for performance prediction.
We propose a novel framework for neural network selection by analyzing the governing dynamics over synaptic connections (edges) during training.
Our framework is built on the fact that back-propagation during neural network training is equivalent to the dynamical evolution of synaptic connections.
arXiv Detail & Related papers (2022-01-11T20:53:15Z) - Parameter Prediction for Unseen Deep Architectures [23.79630072083828]
We study if we can use deep learning to directly predict parameters by exploiting the past knowledge of training other networks.
We propose a hypernetwork that can predict performant parameters in a single forward pass taking a fraction of a second, even on a CPU.
The proposed model achieves surprisingly good performance on unseen and diverse networks.
arXiv Detail & Related papers (2021-10-25T16:52:33Z) - Pruning of Convolutional Neural Networks Using Ising Energy Model [45.4796383952516]
We propose an Ising energy model within an optimization framework for pruning convolutional kernels and hidden units.
Our experiments using ResNets, AlexNet, and SqueezeNet on CIFAR-10 and CIFAR-100 datasets show that the proposed method on average can achieve a pruning rate of more than $50%$ of the trainable parameters.
arXiv Detail & Related papers (2021-02-10T14:00:39Z) - Holistic Filter Pruning for Efficient Deep Neural Networks [25.328005340524825]
"Holistic Filter Pruning" (HFP) is a novel approach for common DNN training that is easy to implement and enables to specify accurate pruning rates.
In various experiments, we give insights into the training and achieve state-of-the-art performance on CIFAR-10 and ImageNet.
arXiv Detail & Related papers (2020-09-17T09:23:36Z) - Modeling from Features: a Mean-field Framework for Over-parameterized
Deep Neural Networks [54.27962244835622]
This paper proposes a new mean-field framework for over- parameterized deep neural networks (DNNs)
In this framework, a DNN is represented by probability measures and functions over its features in the continuous limit.
We illustrate the framework via the standard DNN and the Residual Network (Res-Net) architectures.
arXiv Detail & Related papers (2020-07-03T01:37:16Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.