Accelerating Evolutionary Neural Architecture Search via Multi-Fidelity
Evaluation
- URL: http://arxiv.org/abs/2108.04541v1
- Date: Tue, 10 Aug 2021 09:32:26 GMT
- Title: Accelerating Evolutionary Neural Architecture Search via Multi-Fidelity
Evaluation
- Authors: Shangshang Yang, Ye Tian, Xiaoshu Xiang, Shichen Peng, and Xingyi
Zhang
- Abstract summary: We propose an accelerated ENAS via multifidelity evaluation termed MFENAS.
MFENAS achieves a 2.39% test error rate at the cost of only 0.6 GPU days on one NVIDIA 2080TI GPU.
Results on CIFAR-10 show that the architecture obtained by the proposed MFENAS achieves a 2.39% test error rate at the cost of only 0.6 GPU days.
- Score: 5.754705118117044
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Evolutionary neural architecture search (ENAS) has recently received
increasing attention by effectively finding high-quality neural architectures,
which however consumes high computational cost by training the architecture
encoded by each individual for complete epochs in individual evaluation.
Numerous ENAS approaches have been developed to reduce the evaluation cost, but
it is often difficult for most of these approaches to achieve high evaluation
accuracy. To address this issue, in this paper we propose an accelerated ENAS
via multifidelity evaluation termed MFENAS, where the individual evaluation
cost is significantly reduced by training the architecture encoded by each
individual for only a small number of epochs. The balance between evaluation
cost and evaluation accuracy is well maintained by suggesting a multi-fidelity
evaluation, which identifies the potentially good individuals that cannot
survive from previous generations by integrating multiple evaluations under
different numbers of training epochs. For high diversity of neural
architectures, a population initialization strategy is devised to produce
different neural architectures varying from ResNet-like architectures to
Inception-like ones. Experimental results on CIFAR-10 show that the
architecture obtained by the proposed MFENAS achieves a 2.39% test error rate
at the cost of only 0.6 GPU days on one NVIDIA 2080TI GPU, demonstrating the
superiority of the proposed MFENAS over state-of-the-art NAS approaches in
terms of both computational cost and architecture quality. The architecture
obtained by the proposed MFENAS is then transferred to CIFAR-100 and ImageNet,
which also exhibits competitive performance to the architectures obtained by
existing NAS approaches. The source code of the proposed MFENAS is available at
https://github.com/DevilYangS/MFENAS/.
Related papers
- A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - DNA Family: Boosting Weight-Sharing NAS with Block-Wise Supervisions [121.05720140641189]
We develop a family of models with the distilling neural architecture (DNA) techniques.
Our proposed DNA models can rate all architecture candidates, as opposed to previous works that can only access a sub- search space using algorithms.
Our models achieve state-of-the-art top-1 accuracy of 78.9% and 83.6% on ImageNet for a mobile convolutional network and a small vision transformer, respectively.
arXiv Detail & Related papers (2024-03-02T22:16:47Z) - Efficient Search of Comprehensively Robust Neural Architectures via
Multi-fidelity Evaluation [1.9100854225243937]
We propose a novel efficient search of comprehensively robust neural architectures via multi-fidelity evaluation (ES-CRNA-ME)
Specifically, we first search for comprehensively robust architectures under multiple types of evaluations using the weight-sharing-based NAS method.
We reduce the number of robustness evaluations by the correlation analysis, which can incorporate similar evaluations and decrease the evaluation cost.
arXiv Detail & Related papers (2023-05-12T08:28:58Z) - PRE-NAS: Predictor-assisted Evolutionary Neural Architecture Search [34.06028035262884]
We propose a novel evolutionary-based NAS strategy, Predictor-assisted E-NAS (PRE-NAS)
PRE-NAS leverages new evolutionary search strategies and integrates high-fidelity weight inheritance over generations.
Experiments on NAS-Bench-201 and DARTS search spaces show that PRE-NAS can outperform state-of-the-art NAS methods.
arXiv Detail & Related papers (2022-04-27T06:40:39Z) - Homogeneous Architecture Augmentation for Neural Predictor [13.35821898997164]
Neural Architecture Search (NAS) can automatically design well-performed architectures of Deep Neural Networks (DNNs) for the tasks at hand.
One bottleneck of NAS is the computational cost largely due to the expensive performance evaluation.
Despite their popularity, they also suffer a severe limitation: the shortage of annotated DNN architectures for effectively training the neural predictors.
arXiv Detail & Related papers (2021-07-28T03:46:33Z) - Weak NAS Predictors Are All You Need [91.11570424233709]
Recent predictor-based NAS approaches attempt to solve the problem with two key steps: sampling some architecture-performance pairs and fitting a proxy accuracy predictor.
We shift the paradigm from finding a complicated predictor that covers the whole architecture space to a set of weaker predictors that progressively move towards the high-performance sub-space.
Our method costs fewer samples to find the top-performance architectures on NAS-Bench-101 and NAS-Bench-201, and it achieves the state-of-the-art ImageNet performance on the NASNet search space.
arXiv Detail & Related papers (2021-02-21T01:58:43Z) - Multi-objective Neural Architecture Search with Almost No Training [9.93048700248444]
We propose an effective alternative, dubbed Random-Weight Evaluation (RWE), to rapidly estimate the performance of network architectures.
RWE reduces the computational cost of evaluating an architecture from hours to seconds.
When integrated within an evolutionary multi-objective algorithm, RWE obtains a set of efficient architectures with state-of-the-art performance on CIFAR-10 with less than two hours' searching on a single GPU card.
arXiv Detail & Related papers (2020-11-27T07:39:17Z) - Binarized Neural Architecture Search for Efficient Object Recognition [120.23378346337311]
Binarized neural architecture search (BNAS) produces extremely compressed models to reduce huge computational cost on embedded devices for edge computing.
An accuracy of $96.53%$ vs. $97.22%$ is achieved on the CIFAR-10 dataset, but with a significantly compressed model, and a $40%$ faster search than the state-of-the-art PC-DARTS.
arXiv Detail & Related papers (2020-09-08T15:51:23Z) - Multi-fidelity Neural Architecture Search with Knowledge Distillation [69.09782590880367]
We propose a bayesian multi-fidelity method for neural architecture search: MF-KD.
Knowledge distillation adds to a loss function a term forcing a network to mimic some teacher network.
We show that training for a few epochs with such a modified loss function leads to a better selection of neural architectures than training for a few epochs with a logistic loss.
arXiv Detail & Related papers (2020-06-15T12:32:38Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.