DropNAS: Grouped Operation Dropout for Differentiable Architecture
Search
- URL: http://arxiv.org/abs/2201.11679v1
- Date: Thu, 27 Jan 2022 17:28:23 GMT
- Title: DropNAS: Grouped Operation Dropout for Differentiable Architecture
Search
- Authors: Weijun Hong, Guilin Li, Weinan Zhang, Ruiming Tang, Yunhe Wang,
Zhenguo Li, Yong Yu
- Abstract summary: Recently, DARTS relaxes the search process with a differentiable formulation that leverages weight-sharing and SGD.
This causes two problems: firstly, the operations with more parameters may never have the chance to express the desired function.
We propose a novel grouped operation dropout algorithm named DropNAS to fix the problems with DARTS.
- Score: 78.06809383150437
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural architecture search (NAS) has shown encouraging results in automating
the architecture design. Recently, DARTS relaxes the search process with a
differentiable formulation that leverages weight-sharing and SGD where all
candidate operations are trained simultaneously. Our empirical results show
that such procedure results in the co-adaption problem and Matthew Effect:
operations with fewer parameters would be trained maturely earlier. This causes
two problems: firstly, the operations with more parameters may never have the
chance to express the desired function since those with less have already done
the job; secondly, the system will punish those underperforming operations by
lowering their architecture parameter, and they will get smaller loss
gradients, which causes the Matthew Effect. In this paper, we systematically
study these problems and propose a novel grouped operation dropout algorithm
named DropNAS to fix the problems with DARTS. Extensive experiments demonstrate
that DropNAS solves the above issues and achieves promising performance.
Specifically, DropNAS achieves 2.26% test error on CIFAR-10, 16.39% on
CIFAR-100 and 23.4% on ImageNet (with the same training hyperparameters as
DARTS for a fair comparison). It is also observed that DropNAS is robust across
variants of the DARTS search space. Code is available at
https://github.com/wiljohnhong/DropNAS.
Related papers
- $\beta$-DARTS: Beta-Decay Regularization for Differentiable Architecture
Search [85.84110365657455]
We propose a simple-but-efficient regularization method, termed as Beta-Decay, to regularize the DARTS-based NAS searching process.
Experimental results on NAS-Bench-201 show that our proposed method can help to stabilize the searching process and makes the searched network more transferable across different datasets.
arXiv Detail & Related papers (2022-03-03T11:47:14Z) - L$^{2}$NAS: Learning to Optimize Neural Architectures via
Continuous-Action Reinforcement Learning [23.25155249879658]
Differentiable architecture search (NAS) achieved remarkable results in deep neural network design.
We show that L$2$ achieves state-of-theart results on DART201 benchmark as well as NASS and Once-for-All search policies.
arXiv Detail & Related papers (2021-09-25T19:26:30Z) - Pi-NAS: Improving Neural Architecture Search by Reducing Supernet
Training Consistency Shift [128.32670289503025]
Recently proposed neural architecture search (NAS) methods co-train billions of architectures in a supernet and estimate their potential accuracy.
The ranking correlation between the architectures' predicted accuracy and their actual capability is incorrect, which causes the existing NAS methods' dilemma.
We attribute this ranking correlation problem to the supernet training consistency shift, including feature shift and parameter shift.
We address these two shifts simultaneously using a nontrivial supernet-Pi model, called Pi-NAS.
arXiv Detail & Related papers (2021-08-22T09:08:48Z) - Zero-Cost Proxies Meet Differentiable Architecture Search [20.957570100784988]
Differentiable neural architecture search (NAS) has attracted significant attention in recent years.
Despite its success, DARTS lacks robustness in certain cases.
We propose a novel operation selection paradigm in the context of differentiable NAS.
arXiv Detail & Related papers (2021-06-12T15:33:36Z) - Binarized Neural Architecture Search for Efficient Object Recognition [120.23378346337311]
Binarized neural architecture search (BNAS) produces extremely compressed models to reduce huge computational cost on embedded devices for edge computing.
An accuracy of $96.53%$ vs. $97.22%$ is achieved on the CIFAR-10 dataset, but with a significantly compressed model, and a $40%$ faster search than the state-of-the-art PC-DARTS.
arXiv Detail & Related papers (2020-09-08T15:51:23Z) - DA-NAS: Data Adapted Pruning for Efficient Neural Architecture Search [76.9225014200746]
Efficient search is a core issue in Neural Architecture Search (NAS)
We present DA-NAS that can directly search the architecture for large-scale target tasks while allowing a large candidate set in a more efficient manner.
It is 2x faster than previous methods while the accuracy is currently state-of-the-art, at 76.2% under small FLOPs constraint.
arXiv Detail & Related papers (2020-03-27T17:55:21Z) - DSNAS: Direct Neural Architecture Search without Parameter Retraining [112.02966105995641]
We propose a new problem definition for NAS, task-specific end-to-end, based on this observation.
We propose DSNAS, an efficient differentiable NAS framework that simultaneously optimize architecture and parameters with a low-biased Monte Carlo estimate.
DSNAS successfully discovers networks with comparable accuracy (74.4%) on ImageNet in 420 GPU hours, reducing the total time by more than 34%.
arXiv Detail & Related papers (2020-02-21T04:41:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.