Semi-Supervised Neural Architecture Search
- URL: http://arxiv.org/abs/2002.10389v4
- Date: Tue, 3 Nov 2020 09:44:09 GMT
- Title: Semi-Supervised Neural Architecture Search
- Authors: Renqian Luo, Xu Tan, Rui Wang, Tao Qin, Enhong Chen, Tie-Yan Liu
- Abstract summary: SemiNAS is a semi-supervised Neural architecture search (NAS) approach that leverages numerous unlabeled architectures (without evaluation and thus nearly no cost)
It achieves 94.02% test accuracy on NASBench-101, outperforming all the baselines when using the same number of architectures.
It achieves 97% intelligibility rate in the low-resource setting and 15% test error rate in the robustness setting, with 9%, 7% improvements over the baseline respectively.
- Score: 185.0651567642238
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural architecture search (NAS) relies on a good controller to generate
better architectures or predict the accuracy of given architectures. However,
training the controller requires both abundant and high-quality pairs of
architectures and their accuracy, while it is costly to evaluate an
architecture and obtain its accuracy. In this paper, we propose SemiNAS, a
semi-supervised NAS approach that leverages numerous unlabeled architectures
(without evaluation and thus nearly no cost). Specifically, SemiNAS 1) trains
an initial accuracy predictor with a small set of architecture-accuracy data
pairs; 2) uses the trained accuracy predictor to predict the accuracy of large
amount of architectures (without evaluation); and 3) adds the generated data
pairs to the original data to further improve the predictor. The trained
accuracy predictor can be applied to various NAS algorithms by predicting the
accuracy of candidate architectures for them. SemiNAS has two advantages: 1) It
reduces the computational cost under the same accuracy guarantee. On
NASBench-101 benchmark dataset, it achieves comparable accuracy with
gradient-based method while using only 1/7 architecture-accuracy pairs. 2) It
achieves higher accuracy under the same computational cost. It achieves 94.02%
test accuracy on NASBench-101, outperforming all the baselines when using the
same number of architectures. On ImageNet, it achieves 23.5% top-1 error rate
(under 600M FLOPS constraint) using 4 GPU-days for search. We further apply it
to LJSpeech text to speech task and it achieves 97% intelligibility rate in the
low-resource setting and 15% test error rate in the robustness setting, with
9%, 7% improvements over the baseline respectively.
Related papers
- DAS: Neural Architecture Search via Distinguishing Activation Score [21.711985665733653]
Neural Architecture Search (NAS) is an automatic technique that can search for well-performed architectures for a specific task.
We propose a dataset called Darts-training-bench (DTB), which fills the gap that no training states of architecture in existing datasets.
Our proposed method has 1.04$times$ - 1.56$times$ improvements on NAS-Bench-101, Network Design Spaces, and the proposed DTB.
arXiv Detail & Related papers (2022-12-23T04:02:46Z) - NAAP-440 Dataset and Baseline for Neural Architecture Accuracy
Prediction [1.2183405753834562]
We introduce the NAAP-440 dataset of 440 neural architectures, which were trained on CIFAR10 using a fixed recipe.
Experiments indicate that by using off-the-shelf regression algorithms and running up to 10% of the training process, not only is it possible to predict an architecture's accuracy rather precisely.
This approach may serve as a powerful tool for accelerating NAS-based studies and thus dramatically increase their efficiency.
arXiv Detail & Related papers (2022-09-14T13:21:39Z) - BaLeNAS: Differentiable Architecture Search via the Bayesian Learning
Rule [95.56873042777316]
Differentiable Architecture Search (DARTS) has received massive attention in recent years, mainly because it significantly reduces the computational cost.
This paper formulates the neural architecture search as a distribution learning problem through relaxing the architecture weights into Gaussian distributions.
We demonstrate how the differentiable NAS benefits from Bayesian principles, enhancing exploration and improving stability.
arXiv Detail & Related papers (2021-11-25T18:13:42Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - AceNAS: Learning to Rank Ace Neural Architectures with Weak Supervision
of Weight Sharing [6.171090327531059]
We introduce Learning to Rank methods to select the best (ace) architectures from a space.
We also propose to leverage weak supervision from weight sharing by pretraining architecture representation on weak labels obtained from the super-net.
Experiments on NAS benchmarks and large-scale search spaces demonstrate that our approach outperforms SOTA with a significantly reduced search cost.
arXiv Detail & Related papers (2021-08-06T08:31:42Z) - Weak NAS Predictors Are All You Need [91.11570424233709]
Recent predictor-based NAS approaches attempt to solve the problem with two key steps: sampling some architecture-performance pairs and fitting a proxy accuracy predictor.
We shift the paradigm from finding a complicated predictor that covers the whole architecture space to a set of weaker predictors that progressively move towards the high-performance sub-space.
Our method costs fewer samples to find the top-performance architectures on NAS-Bench-101 and NAS-Bench-201, and it achieves the state-of-the-art ImageNet performance on the NASNet search space.
arXiv Detail & Related papers (2021-02-21T01:58:43Z) - Accuracy Prediction with Non-neural Model for Neural Architecture Search [185.0651567642238]
We study an alternative approach which uses non-neural model for accuracy prediction.
We leverage gradient boosting decision tree (GBDT) as the predictor for Neural architecture search (NAS)
Experiments on NASBench-101 and ImageNet demonstrate the effectiveness of using GBDT as predictor for NAS.
arXiv Detail & Related papers (2020-07-09T13:28:49Z) - FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining [65.39532971991778]
We present an accuracy predictor that scores architecture and training recipes jointly, guiding both sample selection and ranking.
We run fast evolutionary searches in just CPU minutes to generate architecture-recipe pairs for a variety of resource constraints.
FBNetV3 makes up a family of state-of-the-art compact neural networks that outperform both automatically and manually-designed competitors.
arXiv Detail & Related papers (2020-06-03T05:20:21Z) - PONAS: Progressive One-shot Neural Architecture Search for Very
Efficient Deployment [9.442139459221783]
We propose Progressive One-shot Neural Architecture Search (PONAS) that combines advantages of progressive NAS and one-shot methods.
PONAS is able to find architecture of a specialized network in around 10 seconds.
In ImageNet classification, 75.2% top-1 accuracy can be obtained, which is comparable with the state of the arts.
arXiv Detail & Related papers (2020-03-11T05:00:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.