Cyclic Differentiable Architecture Search
- URL: http://arxiv.org/abs/2006.10724v4
- Date: Mon, 25 Apr 2022 06:20:19 GMT
- Title: Cyclic Differentiable Architecture Search
- Authors: Hongyuan Yu, Houwen Peng, Yan Huang, Jianlong Fu, Hao Du, Liang Wang,
Haibin Ling
- Abstract summary: Differentiable ARchiTecture Search, i.e., DARTS, has drawn great attention in neural architecture search.
We propose new joint objectives and a novel Cyclic Differentiable ARchiTecture Search framework, dubbed CDARTS.
In the DARTS search space, we achieve 97.52% top-1 accuracy on CIFAR10 and 76.3% top-1 accuracy on ImageNet.
- Score: 99.12381460261841
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Differentiable ARchiTecture Search, i.e., DARTS, has drawn great attention in
neural architecture search. It tries to find the optimal architecture in a
shallow search network and then measures its performance in a deep evaluation
network. The independent optimization of the search and evaluation networks,
however, leaves room for potential improvement by allowing interaction between
the two networks. To address the problematic optimization issue, we propose new
joint optimization objectives and a novel Cyclic Differentiable ARchiTecture
Search framework, dubbed CDARTS. Considering the structure difference, CDARTS
builds a cyclic feedback mechanism between the search and evaluation networks
with introspective distillation. First, the search network generates an initial
architecture for evaluation, and the weights of the evaluation network are
optimized. Second, the architecture weights in the search network are further
optimized by the label supervision in classification, as well as the
regularization from the evaluation network through feature distillation.
Repeating the above cycle results in joint optimization of the search and
evaluation networks and thus enables the evolution of the architecture to fit
the final evaluation network. The experiments and analysis on CIFAR, ImageNet
and NAS-Bench-201 demonstrate the effectiveness of the proposed approach over
the state-of-the-art ones. Specifically, in the DARTS search space, we achieve
97.52% top-1 accuracy on CIFAR10 and 76.3% top-1 accuracy on ImageNet. In the
chain-structured search space, we achieve 78.2% top-1 accuracy on ImageNet,
which is 1.1% higher than EfficientNet-B0. Our code and models are publicly
available at https://github.com/microsoft/Cream.
Related papers
- Pushing the Efficiency Limit Using Structured Sparse Convolutions [82.31130122200578]
We propose Structured Sparse Convolution (SSC), which leverages the inherent structure in images to reduce the parameters in the convolutional filter.
We show that SSC is a generalization of commonly used layers (depthwise, groupwise and pointwise convolution) in efficient architectures''
Architectures based on SSC achieve state-of-the-art performance compared to baselines on CIFAR-10, CIFAR-100, Tiny-ImageNet, and ImageNet classification benchmarks.
arXiv Detail & Related papers (2022-10-23T18:37:22Z) - Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - De-IReps: Searching for improved Re-parameterizing Architecture based on
Differentiable Evolution Strategy [5.495046508448319]
We design a search space that covers almost all re- parameterization operations.
In this search space, multiple-path networks can be unconditionally re- parameterized into single-path networks.
We visualize the features of the searched architecture and give our explanation for the appearance of this architecture.
arXiv Detail & Related papers (2022-04-13T14:07:20Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - Single-level Optimization For Differential Architecture Search [6.3531384587183135]
differential architecture search (DARTS) makes gradient of architecture parameters biased for network weights.
We propose to use single-level to replace bi-level optimization and non-competitive activation function like sigmoid to replace softmax.
Experiments on NAS Benchmark 201 validate our hypothesis and stably find out nearly the optimal architecture.
arXiv Detail & Related papers (2020-12-15T18:40:33Z) - ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse
Coding [86.40042104698792]
We formulate neural architecture search as a sparse coding problem.
In experiments, our two-stage method on CIFAR-10 requires only 0.05 GPU-day for search.
Our one-stage method produces state-of-the-art performances on both CIFAR-10 and ImageNet at the cost of only evaluation time.
arXiv Detail & Related papers (2020-10-13T04:34:24Z) - Off-Policy Reinforcement Learning for Efficient and Effective GAN
Architecture Search [50.40004966087121]
We introduce a new reinforcement learning based neural architecture search (NAS) methodology for generative adversarial network (GAN) architecture search.
The key idea is to formulate the GAN architecture search problem as a Markov decision process (MDP) for smoother architecture sampling.
We exploit an off-policy GAN architecture search algorithm that makes efficient use of the samples generated by previous policies.
arXiv Detail & Related papers (2020-07-17T18:29:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.