Evolving Neural Architecture Using One Shot Model
- URL: http://arxiv.org/abs/2012.12540v1
- Date: Wed, 23 Dec 2020 08:40:53 GMT
- Title: Evolving Neural Architecture Using One Shot Model
- Authors: Nilotpal Sinha, Kuan-Wen Chen
- Abstract summary: We propose a novel way of applying a simple genetic algorithm to the NAS problem called EvNAS (Evolving Neural Architecture using One Shot Model)
EvNAS searches for the architecture on the proxy dataset i.e. CIFAR-10 for 4.4 GPU day on a single GPU and achieves top-1 test error of 2.47%.
Results show the potential of evolutionary methods in solving the architecture search problem.
- Score: 5.188825486231326
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Neural Architecture Search (NAS) is emerging as a new research direction
which has the potential to replace the hand-crafted neural architectures
designed for specific tasks. Previous evolution based architecture search
requires high computational resources resulting in high search time. In this
work, we propose a novel way of applying a simple genetic algorithm to the NAS
problem called EvNAS (Evolving Neural Architecture using One Shot Model) which
reduces the search time significantly while still achieving better result than
previous evolution based methods. The architectures are represented by using
the architecture parameter of the one shot model which results in the weight
sharing among the architectures for a given population of architectures and
also weight inheritance from one generation to the next generation of
architectures. We propose a decoding technique for the architecture parameter
which is used to divert majority of the gradient information towards the given
architecture and is also used for improving the performance prediction of the
given architecture from the one shot model during the search process.
Furthermore, we use the accuracy of the partially trained architecture on the
validation data as a prediction of its fitness in order to reduce the search
time. EvNAS searches for the architecture on the proxy dataset i.e. CIFAR-10
for 4.4 GPU day on a single GPU and achieves top-1 test error of 2.47% with
3.63M parameters which is then transferred to CIFAR-100 and ImageNet achieving
top-1 error of 16.37% and top-5 error of 7.4% respectively. All of these
results show the potential of evolutionary methods in solving the architecture
search problem.
Related papers
- Knowledge-aware Evolutionary Graph Neural Architecture Search [49.13787973318586]
Graph neural architecture search (GNAS) can customize high-performance graph neural network architectures for specific graph tasks or datasets.
Existing GNAS methods begin searching for architectures from a zero-knowledge state, ignoring the prior knowledge that may improve the search efficiency.
This study proposes exploiting such prior knowledge to accelerate the multi-objective evolutionary search on a new graph dataset.
arXiv Detail & Related papers (2024-11-26T11:32:45Z) - A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - NAAP-440 Dataset and Baseline for Neural Architecture Accuracy
Prediction [1.2183405753834562]
We introduce the NAAP-440 dataset of 440 neural architectures, which were trained on CIFAR10 using a fixed recipe.
Experiments indicate that by using off-the-shelf regression algorithms and running up to 10% of the training process, not only is it possible to predict an architecture's accuracy rather precisely.
This approach may serve as a powerful tool for accelerating NAS-based studies and thus dramatically increase their efficiency.
arXiv Detail & Related papers (2022-09-14T13:21:39Z) - Network Graph Based Neural Architecture Search [57.78724765340237]
We search neural network by rewiring the corresponding graph and predict the architecture performance by graph properties.
Because we do not perform machine learning over the entire graph space, the searching process is remarkably efficient.
arXiv Detail & Related papers (2021-12-15T00:12:03Z) - Rethinking Architecture Selection in Differentiable NAS [74.61723678821049]
Differentiable Neural Architecture Search is one of the most popular NAS methods for its search efficiency and simplicity.
We propose an alternative perturbation-based architecture selection that directly measures each operation's influence on the supernet.
We find that several failure modes of DARTS can be greatly alleviated with the proposed selection method.
arXiv Detail & Related papers (2021-08-10T00:53:39Z) - Neural Architecture Search using Covariance Matrix Adaptation Evolution
Strategy [6.8129169853808795]
We propose a framework of applying the Covariance Matrix Adaptation Evolution Strategy (CMA-ES) to the neural architecture search problem called CMANAS.
The architecture are modelled using a normal distribution, which is updated using CMA-ES based on the fitness of the sampled population.
CMANAS finished the architecture search on CIFAR-10 with the top-1 test accuracy of 97.44% in 0.45 GPU day and on CIFAR-100 with the top-1 test accuracy of 83.24% for 0.6 GPU day on a single GPU.
arXiv Detail & Related papers (2021-07-15T11:41:23Z) - Weak NAS Predictors Are All You Need [91.11570424233709]
Recent predictor-based NAS approaches attempt to solve the problem with two key steps: sampling some architecture-performance pairs and fitting a proxy accuracy predictor.
We shift the paradigm from finding a complicated predictor that covers the whole architecture space to a set of weaker predictors that progressively move towards the high-performance sub-space.
Our method costs fewer samples to find the top-performance architectures on NAS-Bench-101 and NAS-Bench-201, and it achieves the state-of-the-art ImageNet performance on the NASNet search space.
arXiv Detail & Related papers (2021-02-21T01:58:43Z) - Multi-Objective Neural Architecture Search Based on Diverse Structures
and Adaptive Recommendation [4.595675084986132]
The search space of neural architecture search (NAS) for convolutional neural network (CNN) is huge.
We propose MoARR algorithm, which utilizes the existing research results and historical information to quickly find architectures that are both lightweight and accurate.
Experimental results show that our MoARR can achieve a powerful and lightweight model (with 1.9% error rate and 2.3M parameters) on CIFAR-10 in 6 GPU hours.
arXiv Detail & Related papers (2020-07-06T13:42:33Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z) - ADWPNAS: Architecture-Driven Weight Prediction for Neural Architecture
Search [6.458169480971417]
We propose an Architecture-Driven Weight Prediction (ADWP) approach for neural architecture search (NAS)
In our approach, we first design an architecture-intensive search space and then train a HyperNetwork by inputting encoding architecture parameters.
Results show that one search procedure can be completed in 4.0 GPU hours on CIFAR-10.
arXiv Detail & Related papers (2020-03-03T05:06:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.