Grow-Push-Prune: aligning deep discriminants for effective structural
network compression
- URL: http://arxiv.org/abs/2009.13716v3
- Date: Sat, 2 Oct 2021 01:03:05 GMT
- Title: Grow-Push-Prune: aligning deep discriminants for effective structural
network compression
- Authors: Qing Tian, Tal Arbel, James J. Clark
- Abstract summary: This paper attempts to derive task-dependent compact models from a deep discriminant analysis perspective.
We propose an iterative and proactive approach for classification tasks which alternates between a pushing step and a pruning step.
Experiments on the MNIST, CIFAR10, and ImageNet datasets demonstrate our approach's efficacy.
- Score: 5.532477732693
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most of today's popular deep architectures are hand-engineered to be
generalists. However, this design procedure usually leads to massive redundant,
useless, or even harmful features for specific tasks. Unnecessarily high
complexities render deep nets impractical for many real-world applications,
especially those without powerful GPU support. In this paper, we attempt to
derive task-dependent compact models from a deep discriminant analysis
perspective. We propose an iterative and proactive approach for classification
tasks which alternates between (1) a pushing step, with an objective to
simultaneously maximize class separation, penalize co-variances, and push deep
discriminants into alignment with a compact set of neurons, and (2) a pruning
step, which discards less useful or even interfering neurons. Deconvolution is
adopted to reverse 'unimportant' filters' effects and recover useful
contributing sources. A simple network growing strategy based on the basic
Inception module is proposed for challenging tasks requiring larger capacity
than what the base net can offer. Experiments on the MNIST, CIFAR10, and
ImageNet datasets demonstrate our approach's efficacy. On ImageNet, by pushing
and pruning our grown Inception-88 model, we achieve more accurate models than
Inception nets generated during growing, residual nets, and popular compact
nets at similar sizes. We also show that our grown Inception nets (without
hard-coded dimension alignment) clearly outperform residual nets of similar
complexities.
Related papers
- The Simpler The Better: An Entropy-Based Importance Metric To Reduce Neural Networks' Depth [5.869633234882029]
We propose an efficiency strategy that leverages prior knowledge transferred by large models.
Simple but effective, we propose a method relying on an Entropy-bASed Importance mEtRic (EASIER) to reduce the depth of over-parametrized deep neural networks.
arXiv Detail & Related papers (2024-04-27T08:28:25Z) - Iterative Soft Shrinkage Learning for Efficient Image Super-Resolution [91.3781512926942]
Image super-resolution (SR) has witnessed extensive neural network designs from CNN to transformer architectures.
This work investigates the potential of network pruning for super-resolution iteration to take advantage of off-the-shelf network designs and reduce the underlying computational overhead.
We propose a novel Iterative Soft Shrinkage-Percentage (ISS-P) method by optimizing the sparse structure of a randomly network at each and tweaking unimportant weights with a small amount proportional to the magnitude scale on-the-fly.
arXiv Detail & Related papers (2023-03-16T21:06:13Z) - RDRN: Recursively Defined Residual Network for Image Super-Resolution [58.64907136562178]
Deep convolutional neural networks (CNNs) have obtained remarkable performance in single image super-resolution.
We propose a novel network architecture which utilizes attention blocks efficiently.
arXiv Detail & Related papers (2022-11-17T11:06:29Z) - Neural Network Compression by Joint Sparsity Promotion and Redundancy
Reduction [4.9613162734482215]
This paper presents a novel training scheme based on composite constraints that prune redundant filters and minimize their effect on overall network learning via sparsity promotion.
Our tests on several pixel-wise segmentation benchmarks show that the number of neurons and the memory footprint of networks in the test phase are significantly reduced without affecting performance.
arXiv Detail & Related papers (2022-10-14T01:34:49Z) - DDCNet: Deep Dilated Convolutional Neural Network for Dense Prediction [0.0]
A receptive field (ERF) and a higher resolution of spatial features within a network are essential for providing higher-resolution dense estimates.
We present a systemic approach to design network architectures that can provide a larger receptive field while maintaining a higher spatial feature resolution.
arXiv Detail & Related papers (2021-07-09T23:15:34Z) - Convolutional Neural Network Pruning with Structural Redundancy
Reduction [11.381864384054824]
We claim that identifying structural redundancy plays a more essential role than finding unimportant filters.
We propose a network pruning approach that identifies structural redundancy of a CNN and prunes filters in the selected layer(s) with the most redundancy.
arXiv Detail & Related papers (2021-04-08T00:16:24Z) - Generic Perceptual Loss for Modeling Structured Output Dependencies [78.59700528239141]
We show that, what matters is the network structure instead of the trained weights.
We demonstrate that a randomly-weighted deep CNN can be used to model the structured dependencies of outputs.
arXiv Detail & Related papers (2021-03-18T23:56:07Z) - Mixed-Privacy Forgetting in Deep Networks [114.3840147070712]
We show that the influence of a subset of the training samples can be removed from the weights of a network trained on large-scale image classification tasks.
Inspired by real-world applications of forgetting techniques, we introduce a novel notion of forgetting in mixed-privacy setting.
We show that our method allows forgetting without having to trade off the model accuracy.
arXiv Detail & Related papers (2020-12-24T19:34:56Z) - Resolution Adaptive Networks for Efficient Inference [53.04907454606711]
We propose a novel Resolution Adaptive Network (RANet), which is inspired by the intuition that low-resolution representations are sufficient for classifying "easy" inputs.
In RANet, the input images are first routed to a lightweight sub-network that efficiently extracts low-resolution representations.
High-resolution paths in the network maintain the capability to recognize the "hard" samples.
arXiv Detail & Related papers (2020-03-16T16:54:36Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.