TF-NAS: Rethinking Three Search Freedoms of Latency-Constrained
Differentiable Neural Architecture Search
- URL: http://arxiv.org/abs/2008.05314v1
- Date: Wed, 12 Aug 2020 13:44:20 GMT
- Title: TF-NAS: Rethinking Three Search Freedoms of Latency-Constrained
Differentiable Neural Architecture Search
- Authors: Yibo Hu, Xiang Wu, Ran He
- Abstract summary: We propose Three-Freedom NAS (TF-NAS) to achieve both good classification accuracy and precise latency constraint.
Experiments on ImageNet demonstrate the effectiveness of TF-NAS. Particularly, our searched TF-NAS-A obtains 76.9% top-1 accuracy, achieving state-of-the-art results with less latency.
- Score: 85.96350089047398
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the flourish of differentiable neural architecture search (NAS),
automatically searching latency-constrained architectures gives a new
perspective to reduce human labor and expertise. However, the searched
architectures are usually suboptimal in accuracy and may have large jitters
around the target latency. In this paper, we rethink three freedoms of
differentiable NAS, i.e. operation-level, depth-level and width-level, and
propose a novel method, named Three-Freedom NAS (TF-NAS), to achieve both good
classification accuracy and precise latency constraint. For the
operation-level, we present a bi-sampling search algorithm to moderate the
operation collapse. For the depth-level, we introduce a sink-connecting search
space to ensure the mutual exclusion between skip and other candidate
operations, as well as eliminate the architecture redundancy. For the
width-level, we propose an elasticity-scaling strategy that achieves precise
latency constraint in a progressively fine-grained manner. Experiments on
ImageNet demonstrate the effectiveness of TF-NAS. Particularly, our searched
TF-NAS-A obtains 76.9% top-1 accuracy, achieving state-of-the-art results with
less latency. The total search time is only 1.8 days on 1 Titan RTX GPU. Code
is available at https://github.com/AberHu/TF-NAS.
Related papers
- $\beta$-DARTS: Beta-Decay Regularization for Differentiable Architecture
Search [85.84110365657455]
We propose a simple-but-efficient regularization method, termed as Beta-Decay, to regularize the DARTS-based NAS searching process.
Experimental results on NAS-Bench-201 show that our proposed method can help to stabilize the searching process and makes the searched network more transferable across different datasets.
arXiv Detail & Related papers (2022-03-03T11:47:14Z) - Zero-Cost Proxies Meet Differentiable Architecture Search [20.957570100784988]
Differentiable neural architecture search (NAS) has attracted significant attention in recent years.
Despite its success, DARTS lacks robustness in certain cases.
We propose a novel operation selection paradigm in the context of differentiable NAS.
arXiv Detail & Related papers (2021-06-12T15:33:36Z) - FNAS: Uncertainty-Aware Fast Neural Architecture Search [54.49650267859032]
Reinforcement learning (RL)-based neural architecture search (NAS) generally guarantees better convergence yet suffers from the requirement of huge computational resources.
We propose a general pipeline to accelerate the convergence of the rollout process as well as the RL process in NAS.
Experiments on the Mobile Neural Architecture Search (MNAS) search space show the proposed Fast Neural Architecture Search (FNAS) accelerates standard RL-based NAS process by 10x.
arXiv Detail & Related papers (2021-05-25T06:32:52Z) - BossNAS: Exploring Hybrid CNN-transformers with Block-wisely
Self-supervised Neural Architecture Search [100.28980854978768]
We present Block-wisely Self-supervised Neural Architecture Search (BossNAS)
We factorize the search space into blocks and utilize a novel self-supervised training scheme, named ensemble bootstrapping, to train each block separately.
We also present HyTra search space, a fabric-like hybrid CNN-transformer search space with searchable down-sampling positions.
arXiv Detail & Related papers (2021-03-23T10:05:58Z) - Neural Architecture Search on ImageNet in Four GPU Hours: A
Theoretically Inspired Perspective [88.39981851247727]
We propose a novel framework called training-free neural architecture search (TE-NAS)
TE-NAS ranks architectures by analyzing the spectrum of the neural tangent kernel (NTK) and the number of linear regions in the input space.
We show that: (1) these two measurements imply the trainability and expressivity of a neural network; (2) they strongly correlate with the network's test accuracy.
arXiv Detail & Related papers (2021-02-23T07:50:44Z) - S3NAS: Fast NPU-aware Neural Architecture Search Methodology [2.607400740040335]
We present a fast NPU-aware NAS methodology, called S3NAS, to find a CNN architecture with higher accuracy than the existing ones.
We are able to find a network in 3 hours using TPUv3, which shows 82.72% top-1 accuracy on ImageNet with 11.66 ms latency.
arXiv Detail & Related papers (2020-09-04T04:45:50Z) - LC-NAS: Latency Constrained Neural Architecture Search for Point Cloud
Networks [73.78551758828294]
LC-NAS is able to find state-of-the-art architectures for point cloud classification with minimal computational cost.
We show how our searched architectures achieve any desired latency with a reasonably low drop in accuracy.
arXiv Detail & Related papers (2020-08-24T10:30:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.