EmProx: Neural Network Performance Estimation For Neural Architecture
Search
- URL: http://arxiv.org/abs/2206.05972v1
- Date: Mon, 13 Jun 2022 08:35:52 GMT
- Title: EmProx: Neural Network Performance Estimation For Neural Architecture
Search
- Authors: G.G.H. Franken, P. Singh, J. Vanschoren
- Abstract summary: This study proposes a new method, EmProx Score (Embedding Proximity Score) to map architectures to a continuous embedding space.
Performance of candidates is then estimated using weighted kNN based on the embedding vectors of architectures of which the performance is known.
Performance estimations of this method are comparable to the performance predictor used in NAO in terms of accuracy, while being nearly nine times faster to train compared to NAO.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Common Neural Architecture Search methods generate large amounts of candidate
architectures that need training in order to assess their performance and find
an optimal architecture. To minimize the search time we use different
performance estimation strategies. The effectiveness of such strategies varies
in terms of accuracy and fit and query time. This study proposes a new method,
EmProx Score (Embedding Proximity Score). Similar to Neural Architecture
Optimization (NAO), this method maps candidate architectures to a continuous
embedding space using an encoder-decoder framework. The performance of
candidates is then estimated using weighted kNN based on the embedding vectors
of architectures of which the performance is known. Performance estimations of
this method are comparable to the MLP performance predictor used in NAO in
terms of accuracy, while being nearly nine times faster to train compared to
NAO. Benchmarking against other performance estimation strategies currently
used shows similar to better accuracy, while being five up to eighty times
faster.
Related papers
- NAAP-440 Dataset and Baseline for Neural Architecture Accuracy
Prediction [1.2183405753834562]
We introduce the NAAP-440 dataset of 440 neural architectures, which were trained on CIFAR10 using a fixed recipe.
Experiments indicate that by using off-the-shelf regression algorithms and running up to 10% of the training process, not only is it possible to predict an architecture's accuracy rather precisely.
This approach may serve as a powerful tool for accelerating NAS-based studies and thus dramatically increase their efficiency.
arXiv Detail & Related papers (2022-09-14T13:21:39Z) - Neural Architecture Search for Speech Emotion Recognition [72.1966266171951]
We propose to apply neural architecture search (NAS) techniques to automatically configure the SER models.
We show that NAS can improve SER performance (54.89% to 56.28%) while maintaining model parameter sizes.
arXiv Detail & Related papers (2022-03-31T10:16:10Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - RANK-NOSH: Efficient Predictor-Based Architecture Search via Non-Uniform
Successive Halving [74.61723678821049]
We propose NOn-uniform Successive Halving (NOSH), a hierarchical scheduling algorithm that terminates the training of underperforming architectures early to avoid wasting budget.
We formulate predictor-based architecture search as learning to rank with pairwise comparisons.
The resulting method - RANK-NOSH, reduces the search budget by 5x while achieving competitive or even better performance than previous state-of-the-art predictor-based methods on various spaces and datasets.
arXiv Detail & Related papers (2021-08-18T07:45:21Z) - How Powerful are Performance Predictors in Neural Architecture Search? [43.86743225322636]
We give the first large-scale study of performance predictors by analyzing 31 techniques.
We show that certain families of predictors can be combined to achieve even better predictive power.
arXiv Detail & Related papers (2021-04-02T17:57:16Z) - Contrastive Neural Architecture Search with Neural Architecture
Comparators [46.45102111497492]
One of the key steps in Neural Architecture Search (NAS) is to estimate the performance of candidate architectures.
Existing methods either directly use the validation performance or learn a predictor to estimate the performance.
We propose a novel Contrastive Neural Architecture Search (CTNAS) method which performs architecture search by taking the comparison results between architectures as the reward.
arXiv Detail & Related papers (2021-03-08T11:24:07Z) - Off-Policy Reinforcement Learning for Efficient and Effective GAN
Architecture Search [50.40004966087121]
We introduce a new reinforcement learning based neural architecture search (NAS) methodology for generative adversarial network (GAN) architecture search.
The key idea is to formulate the GAN architecture search problem as a Markov decision process (MDP) for smoother architecture sampling.
We exploit an off-policy GAN architecture search algorithm that makes efficient use of the samples generated by previous policies.
arXiv Detail & Related papers (2020-07-17T18:29:17Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z) - Bayesian Neural Architecture Search using A Training-Free Performance
Metric [7.775212462771685]
Recurrent neural networks (RNNs) are a powerful approach for time series prediction.
This paper proposes to tackle the architecture optimization problem with a variant of the Bayesian Optimization (BO) algorithm.
Also, we propose three fixed-length encoding schemes to cope with the variable-length architecture representation.
arXiv Detail & Related papers (2020-01-29T08:42:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.