Tiered Pruning for Efficient Differentialble Inference-Aware Neural
Architecture Search
- URL: http://arxiv.org/abs/2209.11785v1
- Date: Fri, 23 Sep 2022 18:03:54 GMT
- Title: Tiered Pruning for Efficient Differentialble Inference-Aware Neural
Architecture Search
- Authors: S{\l}awomir Kierat, Mateusz Sieniawski, Denys Fridman, Chen-Han Yu,
Szymon Migacz, Pawe{\l} Morkisz, Alex-Fit Florea
- Abstract summary: We introduce, a bi-path building block for DNAS, which can search over inner hidden dimensions with memory and compute complexity.
Second, we present an algorithm for pruning blocks within a layer of the SuperNet during the search.
Third, we describe a novel technique for pruning unnecessary layers during the search.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose three novel pruning techniques to improve the cost and results of
inference-aware Differentiable Neural Architecture Search (DNAS). First, we
introduce , a stochastic bi-path building block for DNAS, which can search over
inner hidden dimensions with memory and compute complexity. Second, we present
an algorithm for pruning blocks within a stochastic layer of the SuperNet
during the search. Third, we describe a novel technique for pruning unnecessary
stochastic layers during the search. The optimized models resulting from the
search are called PruNet and establishes a new state-of-the-art Pareto frontier
for NVIDIA V100 in terms of inference latency for ImageNet Top-1 image
classification accuracy. PruNet as a backbone also outperforms GPUNet and
EfficientNet on the COCO object detection task on inference latency relative to
mean Average Precision (mAP).
Related papers
- Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - Fast Neural Architecture Search for Lightweight Dense Prediction
Networks [41.605107921584775]
We present LDP, a lightweight dense prediction neural architecture search (NAS) framework.
Starting from a pre-defined generic backbone, LDP applies the novel Assisted Tabu Search for efficient architecture exploration.
Experiments show that the proposed framework yields consistent improvements on all tested dense prediction tasks.
arXiv Detail & Related papers (2022-03-03T20:17:10Z) - Combined Depth Space based Architecture Search For Person
Re-identification [70.86236888223569]
We aim to design a lightweight and suitable network for person re-identification (ReID)
We propose a novel search space called Combined Depth Space (CDS), based on which we search for an efficient network architecture, which we call CDNet.
We then propose a low-cost search strategy named the Top-k Sample Search strategy to make full use of the search space and avoid trapping in local optimal result.
arXiv Detail & Related papers (2021-04-09T02:40:01Z) - OPANAS: One-Shot Path Aggregation Network Architecture Search for Object
Detection [82.04372532783931]
Recently, neural architecture search (NAS) has been exploited to design feature pyramid networks (FPNs)
We propose a novel One-Shot Path Aggregation Network Architecture Search (OPANAS) algorithm, which significantly improves both searching efficiency and detection accuracy.
arXiv Detail & Related papers (2021-03-08T01:48:53Z) - Trilevel Neural Architecture Search for Efficient Single Image
Super-Resolution [127.92235484598811]
This paper proposes a trilevel neural architecture search (NAS) method for efficient single image super-resolution (SR)
For modeling the discrete search space, we apply a new continuous relaxation on the discrete search spaces to build a hierarchical mixture of network-path, cell-operations, and kernel-width.
An efficient search algorithm is proposed to perform optimization in a hierarchical supernet manner.
arXiv Detail & Related papers (2021-01-17T12:19:49Z) - ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse
Coding [86.40042104698792]
We formulate neural architecture search as a sparse coding problem.
In experiments, our two-stage method on CIFAR-10 requires only 0.05 GPU-day for search.
Our one-stage method produces state-of-the-art performances on both CIFAR-10 and ImageNet at the cost of only evaluation time.
arXiv Detail & Related papers (2020-10-13T04:34:24Z) - DrNAS: Dirichlet Neural Architecture Search [88.56953713817545]
We treat the continuously relaxed architecture mixing weight as random variables, modeled by Dirichlet distribution.
With recently developed pathwise derivatives, the Dirichlet parameters can be easily optimized with gradient-based generalization.
To alleviate the large memory consumption of differentiable NAS, we propose a simple yet effective progressive learning scheme.
arXiv Detail & Related papers (2020-06-18T08:23:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.