Contrastive Neural Architecture Search with Neural Architecture
Comparators
- URL: http://arxiv.org/abs/2103.05471v1
- Date: Mon, 8 Mar 2021 11:24:07 GMT
- Title: Contrastive Neural Architecture Search with Neural Architecture
Comparators
- Authors: Yaofo Chen, Yong Guo, Qi Chen, Minli Li, Yaowei Wang, Wei Zeng,
Mingkui Tan
- Abstract summary: One of the key steps in Neural Architecture Search (NAS) is to estimate the performance of candidate architectures.
Existing methods either directly use the validation performance or learn a predictor to estimate the performance.
We propose a novel Contrastive Neural Architecture Search (CTNAS) method which performs architecture search by taking the comparison results between architectures as the reward.
- Score: 46.45102111497492
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One of the key steps in Neural Architecture Search (NAS) is to estimate the
performance of candidate architectures. Existing methods either directly use
the validation performance or learn a predictor to estimate the performance.
However, these methods can be either computationally expensive or very
inaccurate, which may severely affect the search efficiency and performance.
Moreover, as it is very difficult to annotate architectures with accurate
performance on specific tasks, learning a promising performance predictor is
often non-trivial due to the lack of labeled data. In this paper, we argue that
it may not be necessary to estimate the absolute performance for NAS. On the
contrary, we may need only to understand whether an architecture is better than
a baseline one. However, how to exploit this comparison information as the
reward and how to well use the limited labeled data remains two great
challenges. In this paper, we propose a novel Contrastive Neural Architecture
Search (CTNAS) method which performs architecture search by taking the
comparison results between architectures as the reward. Specifically, we design
and learn a Neural Architecture Comparator (NAC) to compute the probability of
candidate architectures being better than a baseline one. Moreover, we present
a baseline updating scheme to improve the baseline iteratively in a curriculum
learning manner. More critically, we theoretically show that learning NAC is
equivalent to optimizing the ranking over architectures. Extensive experiments
in three search spaces demonstrate the superiority of our CTNAS over existing
methods.
Related papers
- DAS: Neural Architecture Search via Distinguishing Activation Score [21.711985665733653]
Neural Architecture Search (NAS) is an automatic technique that can search for well-performed architectures for a specific task.
We propose a dataset called Darts-training-bench (DTB), which fills the gap that no training states of architecture in existing datasets.
Our proposed method has 1.04$times$ - 1.56$times$ improvements on NAS-Bench-101, Network Design Spaces, and the proposed DTB.
arXiv Detail & Related papers (2022-12-23T04:02:46Z) - Neural Architecture Ranker [19.21631623578852]
Architecture ranking has recently been advocated to design an efficient and effective performance predictor for Neural Architecture Search (NAS)
Inspired by the stratification stratification, we propose a predictor, namely Neural Ranker (NAR)
arXiv Detail & Related papers (2022-01-30T04:54:59Z) - BaLeNAS: Differentiable Architecture Search via the Bayesian Learning
Rule [95.56873042777316]
Differentiable Architecture Search (DARTS) has received massive attention in recent years, mainly because it significantly reduces the computational cost.
This paper formulates the neural architecture search as a distribution learning problem through relaxing the architecture weights into Gaussian distributions.
We demonstrate how the differentiable NAS benefits from Bayesian principles, enhancing exploration and improving stability.
arXiv Detail & Related papers (2021-11-25T18:13:42Z) - RankNAS: Efficient Neural Architecture Search by Pairwise Ranking [30.890612901949307]
We propose a performance ranking method (RankNAS) via pairwise ranking.
It enables efficient architecture search using much fewer training examples.
It can design high-performance architectures while being orders of magnitude faster than state-of-the-art NAS systems.
arXiv Detail & Related papers (2021-09-15T15:43:08Z) - The Nonlinearity Coefficient -- A Practical Guide to Neural Architecture
Design [3.04585143845864]
We develop methods that can predict, without any training, whether an architecture will achieve a relatively high test or training error on a task after training.
We then go on to explain the error in terms of the architecture definition itself and develop tools for modifying the architecture.
Our first major contribution is to show that the 'degree of nonlinearity' of a neural architecture is a key causal driver behind its performance.
arXiv Detail & Related papers (2021-05-25T20:47:43Z) - Weak NAS Predictors Are All You Need [91.11570424233709]
Recent predictor-based NAS approaches attempt to solve the problem with two key steps: sampling some architecture-performance pairs and fitting a proxy accuracy predictor.
We shift the paradigm from finding a complicated predictor that covers the whole architecture space to a set of weaker predictors that progressively move towards the high-performance sub-space.
Our method costs fewer samples to find the top-performance architectures on NAS-Bench-101 and NAS-Bench-201, and it achieves the state-of-the-art ImageNet performance on the NASNet search space.
arXiv Detail & Related papers (2021-02-21T01:58:43Z) - Does Unsupervised Architecture Representation Learning Help Neural
Architecture Search? [22.63641173256389]
Existing Neural Architecture Search (NAS) methods either encode neural architectures using discrete encodings that do not scale well, or adopt supervised learning-based methods to jointly learn architecture representations and optimize architecture search on such representations which incurs search bias.
We observe that the structural properties of neural architectures are hard to preserve in the latent space if architecture representation learning and search are coupled, resulting in less effective search performance.
arXiv Detail & Related papers (2020-06-12T04:15:34Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z) - Stage-Wise Neural Architecture Search [65.03109178056937]
Modern convolutional networks such as ResNet and NASNet have achieved state-of-the-art results in many computer vision applications.
These networks consist of stages, which are sets of layers that operate on representations in the same resolution.
It has been demonstrated that increasing the number of layers in each stage improves the prediction ability of the network.
However, the resulting architecture becomes computationally expensive in terms of floating point operations, memory requirements and inference time.
arXiv Detail & Related papers (2020-04-23T14:16:39Z) - DA-NAS: Data Adapted Pruning for Efficient Neural Architecture Search [76.9225014200746]
Efficient search is a core issue in Neural Architecture Search (NAS)
We present DA-NAS that can directly search the architecture for large-scale target tasks while allowing a large candidate set in a more efficient manner.
It is 2x faster than previous methods while the accuracy is currently state-of-the-art, at 76.2% under small FLOPs constraint.
arXiv Detail & Related papers (2020-03-27T17:55:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.