Differentiable Architecture Search with Random Features
- URL: http://arxiv.org/abs/2208.08835v1
- Date: Thu, 18 Aug 2022 13:55:27 GMT
- Title: Differentiable Architecture Search with Random Features
- Authors: Xuanyang Zhang, Yonggang Li, Xiangyu Zhang, Yongtao Wang, Jian Sun
- Abstract summary: Differentiable architecture search (DARTS) has significantly promoted the development of NAS techniques because of its high search efficiency and effectiveness but suffers from performance collapse.
In this paper, we make efforts to alleviate the performance collapse problem for DARTS with only training BatchNorm.
- Score: 80.31916993541513
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Differentiable architecture search (DARTS) has significantly promoted the
development of NAS techniques because of its high search efficiency and
effectiveness but suffers from performance collapse. In this paper, we make
efforts to alleviate the performance collapse problem for DARTS from two
aspects. First, we investigate the expressive power of the supernet in DARTS
and then derive a new setup of DARTS paradigm with only training BatchNorm.
Second, we theoretically find that random features dilute the auxiliary
connection role of skip-connection in supernet optimization and enable search
algorithm focus on fairer operation selection, thereby solving the performance
collapse problem. We instantiate DARTS and PC-DARTS with random features to
build an improved version for each named RF-DARTS and RF-PCDARTS respectively.
Experimental results show that RF-DARTS obtains \textbf{94.36\%} test accuracy
on CIFAR-10 (which is the nearest optimal result in NAS-Bench-201), and
achieves the newest state-of-the-art top-1 test error of \textbf{24.0\%} on
ImageNet when transferring from CIFAR-10. Moreover, RF-DARTS performs robustly
across three datasets (CIFAR-10, CIFAR-100, and SVHN) and four search spaces
(S1-S4). Besides, RF-PCDARTS achieves even better results on ImageNet, that is,
\textbf{23.9\%} top-1 and \textbf{7.1\%} top-5 test error, surpassing
representative methods like single-path, training-free, and partial-channel
paradigms directly searched on ImageNet.
Related papers
- OStr-DARTS: Differentiable Neural Architecture Search based on Operation Strength [70.76342136866413]
Differentiable architecture search (DARTS) has emerged as a promising technique for effective neural architecture search.
DARTS suffers from the well-known degeneration issue which can lead to deteriorating architectures.
We propose a novel criterion based on operation strength that estimates the importance of an operation by its effect on the final loss.
arXiv Detail & Related papers (2024-09-22T13:16:07Z) - BaLeNAS: Differentiable Architecture Search via the Bayesian Learning
Rule [95.56873042777316]
Differentiable Architecture Search (DARTS) has received massive attention in recent years, mainly because it significantly reduces the computational cost.
This paper formulates the neural architecture search as a distribution learning problem through relaxing the architecture weights into Gaussian distributions.
We demonstrate how the differentiable NAS benefits from Bayesian principles, enhancing exploration and improving stability.
arXiv Detail & Related papers (2021-11-25T18:13:42Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - MS-DARTS: Mean-Shift Based Differentiable Architecture Search [11.115656548869199]
We propose a Mean-Shift based DARTS (MS-DARTS) to improve stability based on sampling and perturbation.
MS-DARTS archives higher performance over other state-of-the-art NAS methods with reduced search cost.
arXiv Detail & Related papers (2021-08-23T08:06:45Z) - D-DARTS: Distributed Differentiable Architecture Search [75.12821786565318]
Differentiable ARchiTecture Search (DARTS) is one of the most trending Neural Architecture Search (NAS) methods.
We propose D-DARTS, a novel solution that addresses this problem by nesting several neural networks at cell-level.
arXiv Detail & Related papers (2021-08-20T09:07:01Z) - G-DARTS-A: Groups of Channel Parallel Sampling with Attention [22.343277522403742]
We propose an approach named Group-DARTS with Attention (G-DARTS-A) using multiple groups of channels for searching.
Inspired by the partially sampling strategy of PC-DARTS, we use groups channels to sample the super-network to perform a more efficient search.
arXiv Detail & Related papers (2020-10-16T12:58:08Z) - MSR-DARTS: Minimum Stable Rank of Differentiable Architecture Search [45.09936304802425]
In neural architecture search (NAS), differentiable architecture search (DARTS) has recently attracted much attention due to its high efficiency.
We propose a method called minimum stable rank DARTS (MSR-DARTS) for finding a model with the best generalization error.
MSR-DARTS achieves an error rate of 2.54% with 4.0M parameters within 0.3 GPU-days on CIFAR-10, and a top-1 error rate of 23.9% on ImageNet.
arXiv Detail & Related papers (2020-09-19T11:03:39Z) - RARTS: An Efficient First-Order Relaxed Architecture Search Method [5.491655566898372]
Differentiable architecture search (DARTS) is an effective method for data-driven neural network design based on solving a bilevel optimization problem.
We formulate a single level alternative and a relaxed architecture search (RARTS) method that utilizes the whole dataset in architecture learning via both data and network splitting.
For the task of searching topological architecture, i.e., the edges and the operations, RARTS obtains a higher accuracy and 60% reduction of computational cost than second-order DARTS on CIFAR-10.
arXiv Detail & Related papers (2020-08-10T04:55:51Z) - Cyclic Differentiable Architecture Search [99.12381460261841]
Differentiable ARchiTecture Search, i.e., DARTS, has drawn great attention in neural architecture search.
We propose new joint objectives and a novel Cyclic Differentiable ARchiTecture Search framework, dubbed CDARTS.
In the DARTS search space, we achieve 97.52% top-1 accuracy on CIFAR10 and 76.3% top-1 accuracy on ImageNet.
arXiv Detail & Related papers (2020-06-18T17:55:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.