AutoHAS: Efficient Hyperparameter and Architecture Search
- URL: http://arxiv.org/abs/2006.03656v3
- Date: Wed, 7 Apr 2021 06:55:00 GMT
- Title: AutoHAS: Efficient Hyperparameter and Architecture Search
- Authors: Xuanyi Dong, Mingxing Tan, Adams Wei Yu, Daiyi Peng, Bogdan Gabrys,
Quoc V. Le
- Abstract summary: AutoHAS learns to alternately update the shared network weights and a reinforcement learning controller.
A temporary weight is introduced to store the updated weight from the selected HPs.
In experiments, we show AutoHAS is efficient and generalizable to different search spaces, baselines and datasets.
- Score: 104.29883101871083
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Efficient hyperparameter or architecture search methods have shown remarkable
results, but each of them is only applicable to searching for either
hyperparameters (HPs) or architectures. In this work, we propose a unified
pipeline, AutoHAS, to efficiently search for both architectures and
hyperparameters. AutoHAS learns to alternately update the shared network
weights and a reinforcement learning (RL) controller, which learns the
probability distribution for the architecture candidates and HP candidates. A
temporary weight is introduced to store the updated weight from the selected
HPs (by the controller), and a validation accuracy based on this temporary
weight serves as a reward to update the controller. In experiments, we show
AutoHAS is efficient and generalizable to different search spaces, baselines
and datasets. In particular, AutoHAS can improve the accuracy over popular
network architectures, such as ResNet and EfficientNet, on CIFAR-10/100,
ImageNet, and four more other datasets.
Related papers
- ASP: Automatic Selection of Proxy dataset for efficient AutoML [16.813109584129514]
We propose an Automatic Selection of Proxy dataset framework (ASP) to dynamically find the informative proxy subsets of training data at each epoch.
ASP can obtain better results than other data selection methods at all selection ratios.
arXiv Detail & Related papers (2023-10-17T09:36:22Z) - Shapley-NAS: Discovering Operation Contribution for Neural Architecture
Search [96.20505710087392]
We propose a Shapley value based method to evaluate operation contribution (Shapley-NAS) for neural architecture search.
We show that our method outperforms the state-of-the-art methods by a considerable margin with light search cost.
arXiv Detail & Related papers (2022-06-20T14:41:49Z) - DAAS: Differentiable Architecture and Augmentation Policy Search [107.53318939844422]
This work considers the possible coupling between neural architectures and data augmentation and proposes an effective algorithm jointly searching for them.
Our approach achieves 97.91% accuracy on CIFAR-10 and 76.6% Top-1 accuracy on ImageNet dataset, showing the outstanding performance of our search algorithm.
arXiv Detail & Related papers (2021-09-30T17:15:17Z) - Off-Policy Reinforcement Learning for Efficient and Effective GAN
Architecture Search [50.40004966087121]
We introduce a new reinforcement learning based neural architecture search (NAS) methodology for generative adversarial network (GAN) architecture search.
The key idea is to formulate the GAN architecture search problem as a Markov decision process (MDP) for smoother architecture sampling.
We exploit an off-policy GAN architecture search algorithm that makes efficient use of the samples generated by previous policies.
arXiv Detail & Related papers (2020-07-17T18:29:17Z) - Hypernetwork-Based Augmentation [1.6752182911522517]
We propose an efficient gradient-based search algorithm, called Hypernetwork-Based Augmentation (HBA)
Our HBA uses a hypernetwork to approximate a population-based training algorithm.
Our results show that HBA is competitive to the state-of-the-art methods in terms of both search speed and accuracy.
arXiv Detail & Related papers (2020-06-11T10:36:39Z) - FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining [65.39532971991778]
We present an accuracy predictor that scores architecture and training recipes jointly, guiding both sample selection and ranking.
We run fast evolutionary searches in just CPU minutes to generate architecture-recipe pairs for a variety of resource constraints.
FBNetV3 makes up a family of state-of-the-art compact neural networks that outperform both automatically and manually-designed competitors.
arXiv Detail & Related papers (2020-06-03T05:20:21Z) - ADWPNAS: Architecture-Driven Weight Prediction for Neural Architecture
Search [6.458169480971417]
We propose an Architecture-Driven Weight Prediction (ADWP) approach for neural architecture search (NAS)
In our approach, we first design an architecture-intensive search space and then train a HyperNetwork by inputting encoding architecture parameters.
Results show that one search procedure can be completed in 4.0 GPU hours on CIFAR-10.
arXiv Detail & Related papers (2020-03-03T05:06:20Z) - DDPNAS: Efficient Neural Architecture Search via Dynamic Distribution
Pruning [135.27931587381596]
We propose an efficient and unified NAS framework termed DDPNAS via dynamic distribution pruning.
In particular, we first sample architectures from a joint categorical distribution. Then the search space is dynamically pruned and its distribution is updated every few epochs.
With the proposed efficient network generation method, we directly obtain the optimal neural architectures on given constraints.
arXiv Detail & Related papers (2019-05-28T06:35:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.