Single-level Optimization For Differential Architecture Search
- URL: http://arxiv.org/abs/2012.11337v1
- Date: Tue, 15 Dec 2020 18:40:33 GMT
- Title: Single-level Optimization For Differential Architecture Search
- Authors: Pengfei Hou, Ying Jin
- Abstract summary: differential architecture search (DARTS) makes gradient of architecture parameters biased for network weights.
We propose to use single-level to replace bi-level optimization and non-competitive activation function like sigmoid to replace softmax.
Experiments on NAS Benchmark 201 validate our hypothesis and stably find out nearly the optimal architecture.
- Score: 6.3531384587183135
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we point out that differential architecture search (DARTS)
makes gradient of architecture parameters biased for network weights and
architecture parameters are updated in different datasets alternatively in the
bi-level optimization framework. The bias causes the architecture parameters of
non-learnable operations to surpass that of learnable operations. Moreover,
using softmax as architecture parameters' activation function and inappropriate
learning rate would exacerbate the bias. As a result, it's frequently observed
that non-learnable operations are dominated in the search phase. To reduce the
bias, we propose to use single-level to replace bi-level optimization and
non-competitive activation function like sigmoid to replace softmax. As a
result, we could search high-performance architectures steadily. Experiments on
NAS Benchmark 201 validate our hypothesis and stably find out nearly the
optimal architecture. On DARTS space, we search the state-of-the-art
architecture with 77.0% top1 accuracy (training setting follows PDARTS and
without any additional module) on ImageNet-1K and steadily search architectures
up-to 76.5% top1 accuracy (but not select the best from the searched
architectures) which is comparable with current reported best result.
Related papers
- BaLeNAS: Differentiable Architecture Search via the Bayesian Learning
Rule [95.56873042777316]
Differentiable Architecture Search (DARTS) has received massive attention in recent years, mainly because it significantly reduces the computational cost.
This paper formulates the neural architecture search as a distribution learning problem through relaxing the architecture weights into Gaussian distributions.
We demonstrate how the differentiable NAS benefits from Bayesian principles, enhancing exploration and improving stability.
arXiv Detail & Related papers (2021-11-25T18:13:42Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - DAAS: Differentiable Architecture and Augmentation Policy Search [107.53318939844422]
This work considers the possible coupling between neural architectures and data augmentation and proposes an effective algorithm jointly searching for them.
Our approach achieves 97.91% accuracy on CIFAR-10 and 76.6% Top-1 accuracy on ImageNet dataset, showing the outstanding performance of our search algorithm.
arXiv Detail & Related papers (2021-09-30T17:15:17Z) - Single-DARTS: Towards Stable Architecture Search [7.894638544388165]
We propose Single-DARTS, which merely uses single-level optimization, updating network weights and architecture parameters simultaneously with the same data batch.
Experiment results show that Single-DARTS achieves state-of-the-art performance on mainstream search spaces.
arXiv Detail & Related papers (2021-08-18T13:00:39Z) - Rethinking Architecture Selection in Differentiable NAS [74.61723678821049]
Differentiable Neural Architecture Search is one of the most popular NAS methods for its search efficiency and simplicity.
We propose an alternative perturbation-based architecture selection that directly measures each operation's influence on the supernet.
We find that several failure modes of DARTS can be greatly alleviated with the proposed selection method.
arXiv Detail & Related papers (2021-08-10T00:53:39Z) - iDARTS: Differentiable Architecture Search with Stochastic Implicit
Gradients [75.41173109807735]
Differentiable ARchiTecture Search (DARTS) has recently become the mainstream of neural architecture search (NAS)
We tackle the hypergradient computation in DARTS based on the implicit function theorem.
We show that the architecture optimisation with the proposed method, named iDARTS, is expected to converge to a stationary point.
arXiv Detail & Related papers (2021-06-21T00:44:11Z) - Multi-Objective Neural Architecture Search Based on Diverse Structures
and Adaptive Recommendation [4.595675084986132]
The search space of neural architecture search (NAS) for convolutional neural network (CNN) is huge.
We propose MoARR algorithm, which utilizes the existing research results and historical information to quickly find architectures that are both lightweight and accurate.
Experimental results show that our MoARR can achieve a powerful and lightweight model (with 1.9% error rate and 2.3M parameters) on CIFAR-10 in 6 GPU hours.
arXiv Detail & Related papers (2020-07-06T13:42:33Z) - Cyclic Differentiable Architecture Search [99.12381460261841]
Differentiable ARchiTecture Search, i.e., DARTS, has drawn great attention in neural architecture search.
We propose new joint objectives and a novel Cyclic Differentiable ARchiTecture Search framework, dubbed CDARTS.
In the DARTS search space, we achieve 97.52% top-1 accuracy on CIFAR10 and 76.3% top-1 accuracy on ImageNet.
arXiv Detail & Related papers (2020-06-18T17:55:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.