A General-Purpose Transferable Predictor for Neural Architecture Search
- URL: http://arxiv.org/abs/2302.10835v1
- Date: Tue, 21 Feb 2023 17:28:05 GMT
- Title: A General-Purpose Transferable Predictor for Neural Architecture Search
- Authors: Fred X. Han, Keith G. Mills, Fabian Chudak, Parsa Riahi, Mohammad
Salameh, Jialin Zhang, Wei Lu, Shangling Jui, Di Niu
- Abstract summary: We propose a general-purpose neural predictor for Neural Architecture Search (NAS) that can transfer across search spaces.
Experimental results on NAS-Bench-101, 201 and 301 demonstrate the efficacy of our scheme.
- Score: 22.883809911265445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding and modelling the performance of neural architectures is key to
Neural Architecture Search (NAS). Performance predictors have seen widespread
use in low-cost NAS and achieve high ranking correlations between predicted and
ground truth performance in several NAS benchmarks. However, existing
predictors are often designed based on network encodings specific to a
predefined search space and are therefore not generalizable to other search
spaces or new architecture families. In this paper, we propose a
general-purpose neural predictor for NAS that can transfer across search
spaces, by representing any given candidate Convolutional Neural Network (CNN)
with a Computation Graph (CG) that consists of primitive operators. We further
combine our CG network representation with Contrastive Learning (CL) and
propose a graph representation learning procedure that leverages the structural
information of unlabeled architectures from multiple families to train CG
embeddings for our performance predictor. Experimental results on
NAS-Bench-101, 201 and 301 demonstrate the efficacy of our scheme as we achieve
strong positive Spearman Rank Correlation Coefficient (SRCC) on every search
space, outperforming several Zero-Cost Proxies, including Synflow and Jacov,
which are also generalizable predictors across search spaces. Moreover, when
using our proposed general-purpose predictor in an evolutionary neural
architecture search algorithm, we can find high-performance architectures on
NAS-Bench-101 and find a MobileNetV3 architecture that attains 79.2% top-1
accuracy on ImageNet.
Related papers
- CAP: A Context-Aware Neural Predictor for NAS [4.8761456288582945]
We propose a context-aware neural predictor (CAP) which only needs a few annotated architectures for training.
Experimental results in different search spaces demonstrate the superior performance of CAP compared with state-of-the-art neural predictors.
arXiv Detail & Related papers (2024-06-04T07:37:47Z) - FR-NAS: Forward-and-Reverse Graph Predictor for Efficient Neural Architecture Search [10.699485270006601]
We introduce a novel Graph Neural Networks (GNN) predictor for Neural Architecture Search (NAS)
This predictor renders neural architectures into vector representations by combining both the conventional and inverse graph views.
The experimental results showcase a significant improvement in prediction accuracy, with a 3%--16% increase in Kendall-tau correlation.
arXiv Detail & Related papers (2024-04-24T03:22:49Z) - GeNAS: Neural Architecture Search with Better Generalization [14.92869716323226]
Recent neural architecture search (NAS) approaches rely on validation loss or accuracy to find the superior network for the target data.
In this paper, we investigate a new neural architecture search measure for excavating architectures with better generalization.
arXiv Detail & Related papers (2023-05-15T12:44:54Z) - GENNAPE: Towards Generalized Neural Architecture Performance Estimators [25.877126553261434]
GENNAPE represents a given neural network as a Computation Graph (CG) of atomic operations.
It first learns a graph encoder via Contrastive Learning to encourage network separation by topological features.
Experiments show that GENNAPE pretrained on NAS-Bench-101 can achieve superior transferability to 5 different public neural network benchmarks.
arXiv Detail & Related papers (2022-11-30T18:27:41Z) - Landmark Regularization: Ranking Guided Super-Net Training in Neural
Architecture Search [70.57382341642418]
Weight sharing has become a de facto standard in neural architecture search because it enables the search to be done on commodity hardware.
Recent works have empirically shown a ranking disorder between the performance of stand-alone architectures and that of the corresponding shared-weight networks.
We propose a regularization term that aims to maximize the correlation between the performance rankings of the shared-weight network and that of the standalone architectures.
arXiv Detail & Related papers (2021-04-12T09:32:33Z) - Weak NAS Predictors Are All You Need [91.11570424233709]
Recent predictor-based NAS approaches attempt to solve the problem with two key steps: sampling some architecture-performance pairs and fitting a proxy accuracy predictor.
We shift the paradigm from finding a complicated predictor that covers the whole architecture space to a set of weaker predictors that progressively move towards the high-performance sub-space.
Our method costs fewer samples to find the top-performance architectures on NAS-Bench-101 and NAS-Bench-201, and it achieves the state-of-the-art ImageNet performance on the NASNet search space.
arXiv Detail & Related papers (2021-02-21T01:58:43Z) - Hierarchical Neural Architecture Search for Deep Stereo Matching [131.94481111956853]
We propose the first end-to-end hierarchical NAS framework for deep stereo matching.
Our framework incorporates task-specific human knowledge into the neural architecture search framework.
It is ranked at the top 1 accuracy on KITTI stereo 2012, 2015 and Middlebury benchmarks, as well as the top 1 on SceneFlow dataset.
arXiv Detail & Related papers (2020-10-26T11:57:37Z) - Neural Architecture Performance Prediction Using Graph Neural Networks [17.224223176258334]
We propose a surrogate model for neural architecture performance prediction built upon Graph Neural Networks (GNN)
We demonstrate the effectiveness of this surrogate model on neural architecture performance prediction for structurally unknown architectures.
arXiv Detail & Related papers (2020-10-19T09:33:57Z) - FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining [65.39532971991778]
We present an accuracy predictor that scores architecture and training recipes jointly, guiding both sample selection and ranking.
We run fast evolutionary searches in just CPU minutes to generate architecture-recipe pairs for a variety of resource constraints.
FBNetV3 makes up a family of state-of-the-art compact neural networks that outperform both automatically and manually-designed competitors.
arXiv Detail & Related papers (2020-06-03T05:20:21Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.