RED++ : Data-Free Pruning of Deep Neural Networks via Input Splitting
and Output Merging
- URL: http://arxiv.org/abs/2110.01397v1
- Date: Thu, 30 Sep 2021 09:31:11 GMT
- Title: RED++ : Data-Free Pruning of Deep Neural Networks via Input Splitting
and Output Merging
- Authors: Edouard Yvinec, Arnaud Dapogny, Matthieu Cord and Kevin Bailly
- Abstract summary: Pruning Deep Neural Networks (DNNs) is a prominent field of study in the goal of inference runtime acceleration.
We introduce a novel data-free pruning protocol RED++.
We study the theoretical and empirical guarantees on the preservation of the accuracy from the hashing.
- Score: 36.027765880474526
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pruning Deep Neural Networks (DNNs) is a prominent field of study in the goal
of inference runtime acceleration. In this paper, we introduce a novel
data-free pruning protocol RED++. Only requiring a trained neural network, and
not specific to DNN architecture, we exploit an adaptive data-free scalar
hashing which exhibits redundancies among neuron weight values. We study the
theoretical and empirical guarantees on the preservation of the accuracy from
the hashing as well as the expected pruning ratio resulting from the
exploitation of said redundancies. We propose a novel data-free pruning
technique of DNN layers which removes the input-wise redundant operations. This
algorithm is straightforward, parallelizable and offers novel perspective on
DNN pruning by shifting the burden of large computation to efficient memory
access and allocation. We provide theoretical guarantees on RED++ performance
and empirically demonstrate its superiority over other data-free pruning
methods and its competitiveness with data-driven ones on ResNets, MobileNets
and EfficientNets.
Related papers
- Harnessing Neuron Stability to Improve DNN Verification [42.65507402735545]
We present VeriStable, a novel extension of recently proposed DPLL-based constraint DNN verification approach.
We evaluate the effectiveness of VeriStable across a range of challenging benchmarks including fully-connected feed networks (FNNs), convolutional neural networks (CNNs) and residual networks (ResNets)
Preliminary results show that VeriStable is competitive and outperforms state-of-the-art verification tools, including $alpha$-$beta$-CROWN and MN-BaB, the first and second performers of the VNN-COMP, respectively.
arXiv Detail & Related papers (2024-01-19T23:48:04Z) - Deep Neural Networks Tend To Extrapolate Predictably [51.303814412294514]
neural network predictions tend to be unpredictable and overconfident when faced with out-of-distribution (OOD) inputs.
We observe that neural network predictions often tend towards a constant value as input data becomes increasingly OOD.
We show how one can leverage our insights in practice to enable risk-sensitive decision-making in the presence of OOD inputs.
arXiv Detail & Related papers (2023-10-02T03:25:32Z) - Intelligence Processing Units Accelerate Neuromorphic Learning [52.952192990802345]
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy consumption and latency.
We present an IPU-optimized release of our custom SNN Python package, snnTorch.
arXiv Detail & Related papers (2022-11-19T15:44:08Z) - Can pruning improve certified robustness of neural networks? [106.03070538582222]
We show that neural network pruning can improve empirical robustness of deep neural networks (NNs)
Our experiments show that by appropriately pruning an NN, its certified accuracy can be boosted up to 8.2% under standard training.
We additionally observe the existence of certified lottery tickets that can match both standard and certified robust accuracies of the original dense models.
arXiv Detail & Related papers (2022-06-15T05:48:51Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - A Mixed Integer Programming Approach for Verifying Properties of
Binarized Neural Networks [44.44006029119672]
We propose a mixed integer programming formulation for BNN verification.
We demonstrate our approach by verifying properties of BNNs trained on the MNIST dataset and an aircraft collision avoidance controller.
arXiv Detail & Related papers (2022-03-11T01:11:29Z) - RED : Looking for Redundancies for Data-Free Structured Compression of
Deep Neural Networks [36.027765880474526]
We present RED, a data-free structured, unified approach to tackle structured pruning.
First, we propose a novel adaptive hashing of the scalar weight distribution densities to increase the number of identical neurons.
Second, we prune the network by merging redundant neurons based on their relative similarities, as defined by their distance.
Third, we propose a novel uneven depthwise separation technique to further prune convolutional layers.
arXiv Detail & Related papers (2021-05-31T08:44:14Z) - S2-BNN: Bridging the Gap Between Self-Supervised Real and 1-bit Neural
Networks via Guided Distribution Calibration [74.5509794733707]
We present a novel guided learning paradigm from real-valued to distill binary networks on the final prediction distribution.
Our proposed method can boost the simple contrastive learning baseline by an absolute gain of 5.515% on BNNs.
Our method achieves substantial improvement over the simple contrastive learning baseline, and is even comparable to many mainstream supervised BNN methods.
arXiv Detail & Related papers (2021-02-17T18:59:28Z) - Online Limited Memory Neural-Linear Bandits with Likelihood Matching [53.18698496031658]
We study neural-linear bandits for solving problems where both exploration and representation learning play an important role.
We propose a likelihood matching algorithm that is resilient to catastrophic forgetting and is completely online.
arXiv Detail & Related papers (2021-02-07T14:19:07Z) - Encoding the latent posterior of Bayesian Neural Networks for
uncertainty quantification [10.727102755903616]
We aim for efficient deep BNNs amenable to complex computer vision architectures.
We achieve this by leveraging variational autoencoders (VAEs) to learn the interaction and the latent distribution of the parameters at each network layer.
Our approach, Latent-Posterior BNN (LP-BNN), is compatible with the recent BatchEnsemble method, leading to highly efficient (in terms of computation and memory during both training and testing) ensembles.
arXiv Detail & Related papers (2020-12-04T19:50:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.