Growing Efficient Deep Networks by Structured Continuous Sparsification
- URL: http://arxiv.org/abs/2007.15353v2
- Date: Tue, 6 Jun 2023 03:20:50 GMT
- Title: Growing Efficient Deep Networks by Structured Continuous Sparsification
- Authors: Xin Yuan, Pedro Savarese, Michael Maire
- Abstract summary: We develop an approach to growing deep network architectures over the course of training.
Our method can start from a small, simple seed architecture and dynamically grow and prune both layers and filters.
We achieve $49.7%$ inference FLOPs and $47.4%$ training FLOPs savings compared to a baseline ResNet-50 on ImageNet.
- Score: 34.7523496790944
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We develop an approach to growing deep network architectures over the course
of training, driven by a principled combination of accuracy and sparsity
objectives. Unlike existing pruning or architecture search techniques that
operate on full-sized models or supernet architectures, our method can start
from a small, simple seed architecture and dynamically grow and prune both
layers and filters. By combining a continuous relaxation of discrete network
structure optimization with a scheme for sampling sparse subnetworks, we
produce compact, pruned networks, while also drastically reducing the
computational expense of training. For example, we achieve $49.7\%$ inference
FLOPs and $47.4\%$ training FLOPs savings compared to a baseline ResNet-50 on
ImageNet, while maintaining $75.2\%$ top-1 accuracy -- all without any
dedicated fine-tuning stage. Experiments across CIFAR, ImageNet, PASCAL VOC,
and Penn Treebank, with convolutional networks for image classification and
semantic segmentation, and recurrent networks for language modeling,
demonstrate that we both train faster and produce more efficient networks than
competing architecture pruning or search methods.
Related papers
- Network Pruning Spaces [12.692532576302426]
Network pruning techniques, including weight pruning and filter pruning, reveal that most state-of-the-art neural networks can be accelerated without a significant performance drop.
This work focuses on filter pruning which enables accelerated inference with any off-the-shelf deep learning library and hardware.
arXiv Detail & Related papers (2023-04-19T06:52:05Z) - Pushing the Efficiency Limit Using Structured Sparse Convolutions [82.31130122200578]
We propose Structured Sparse Convolution (SSC), which leverages the inherent structure in images to reduce the parameters in the convolutional filter.
We show that SSC is a generalization of commonly used layers (depthwise, groupwise and pointwise convolution) in efficient architectures''
Architectures based on SSC achieve state-of-the-art performance compared to baselines on CIFAR-10, CIFAR-100, Tiny-ImageNet, and ImageNet classification benchmarks.
arXiv Detail & Related papers (2022-10-23T18:37:22Z) - FlowNAS: Neural Architecture Search for Optical Flow Estimation [65.44079917247369]
We propose a neural architecture search method named FlowNAS to automatically find the better encoder architecture for flow estimation task.
Experimental results show that the discovered architecture with the weights inherited from the super-network achieves 4.67% F1-all error on KITTI.
arXiv Detail & Related papers (2022-07-04T09:05:25Z) - Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - Firefly Neural Architecture Descent: a General Approach for Growing
Neural Networks [50.684661759340145]
Firefly neural architecture descent is a general framework for progressively and dynamically growing neural networks.
We show that firefly descent can flexibly grow networks both wider and deeper, and can be applied to learn accurate but resource-efficient neural architectures.
In particular, it learns networks that are smaller in size but have higher average accuracy than those learned by the state-of-the-art methods.
arXiv Detail & Related papers (2021-02-17T04:47:18Z) - Structured Convolutions for Efficient Neural Network Design [65.36569572213027]
We tackle model efficiency by exploiting redundancy in the textitimplicit structure of the building blocks of convolutional neural networks.
We show how this decomposition can be applied to 2D and 3D kernels as well as the fully-connected layers.
arXiv Detail & Related papers (2020-08-06T04:38:38Z) - Paying more attention to snapshots of Iterative Pruning: Improving Model
Compression via Ensemble Distillation [4.254099382808598]
Existing methods often iteratively prune networks to attain high compression ratio without incurring significant loss in performance.
We show that strong ensembles can be constructed from snapshots of iterative pruning, which achieve competitive performance and vary in network structure.
In standard image classification benchmarks such as CIFAR and Tiny-Imagenet, we advance state-of-the-art pruning ratio of structured pruning by integrating simple l1-norm filters pruning into our pipeline.
arXiv Detail & Related papers (2020-06-20T03:59:46Z) - Network Adjustment: Channel Search Guided by FLOPs Utilization Ratio [101.84651388520584]
This paper presents a new framework named network adjustment, which considers network accuracy as a function of FLOPs.
Experiments on standard image classification datasets and a wide range of base networks demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2020-04-06T15:51:00Z) - Knapsack Pruning with Inner Distillation [11.04321604965426]
We propose a novel pruning method that optimize the final accuracy of the pruned network.
We prune the network channels while maintaining the high-level structure of the network.
Our method leads to state-of-the-art pruning results on ImageNet, CIFAR-10 and CIFAR-100 using ResNet backbones.
arXiv Detail & Related papers (2020-02-19T16:04:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.