Architecture Augmentation for Performance Predictor Based on Graph
Isomorphism
- URL: http://arxiv.org/abs/2207.00987v1
- Date: Sun, 3 Jul 2022 09:04:09 GMT
- Title: Architecture Augmentation for Performance Predictor Based on Graph
Isomorphism
- Authors: Xiangning Xie, Yuqiao Liu, Yanan Sun, Mengjie Zhang, Kay Chen Tan
- Abstract summary: We propose an effective deep neural network (DNN) architecture augmentation method named GIAug.
We show that GIAug can significantly enhance the performance of most state-of-the-art peer predictors.
In addition, GIAug can save three magnitude order of computation cost at most on ImageNet.
- Score: 15.478663248038307
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural Architecture Search (NAS) can automatically design architectures for
deep neural networks (DNNs) and has become one of the hottest research topics
in the current machine learning community. However, NAS is often
computationally expensive because a large number of DNNs require to be trained
for obtaining performance during the search process. Performance predictors can
greatly alleviate the prohibitive cost of NAS by directly predicting the
performance of DNNs. However, building satisfactory performance predictors
highly depends on enough trained DNN architectures, which are difficult to
obtain in most scenarios. To solve this critical issue, we propose an effective
DNN architecture augmentation method named GIAug in this paper. Specifically,
we first propose a mechanism based on graph isomorphism, which has the merit of
efficiently generating a factorial of $\boldsymbol n$ (i.e., $\boldsymbol n!$)
diverse annotated architectures upon a single architecture having $\boldsymbol
n$ nodes. In addition, we also design a generic method to encode the
architectures into the form suitable to most prediction models. As a result,
GIAug can be flexibly utilized by various existing performance predictors-based
NAS algorithms. We perform extensive experiments on CIFAR-10 and ImageNet
benchmark datasets on small-, medium- and large-scale search space. The
experiments show that GIAug can significantly enhance the performance of most
state-of-the-art peer predictors. In addition, GIAug can save three magnitude
order of computation cost at most on ImageNet yet with similar performance when
compared with state-of-the-art NAS algorithms.
Related papers
- DNA Family: Boosting Weight-Sharing NAS with Block-Wise Supervisions [121.05720140641189]
We develop a family of models with the distilling neural architecture (DNA) techniques.
Our proposed DNA models can rate all architecture candidates, as opposed to previous works that can only access a sub- search space using algorithms.
Our models achieve state-of-the-art top-1 accuracy of 78.9% and 83.6% on ImageNet for a mobile convolutional network and a small vision transformer, respectively.
arXiv Detail & Related papers (2024-03-02T22:16:47Z) - Do Not Train It: A Linear Neural Architecture Search of Graph Neural
Networks [15.823247346294089]
We develop a novel NAS-GNNs method, namely neural architecture coding (NAC)
Our approach leads to state-of-the-art performance, which is up to $200times$ faster and $18.8%$ more accurate than the strong baselines.
arXiv Detail & Related papers (2023-05-23T13:44:04Z) - GeNAS: Neural Architecture Search with Better Generalization [14.92869716323226]
Recent neural architecture search (NAS) approaches rely on validation loss or accuracy to find the superior network for the target data.
In this paper, we investigate a new neural architecture search measure for excavating architectures with better generalization.
arXiv Detail & Related papers (2023-05-15T12:44:54Z) - GraphPNAS: Learning Distribution of Good Neural Architectures via Deep
Graph Generative Models [48.57083463364353]
We study neural architecture search (NAS) through the lens of learning random graph models.
We propose GraphPNAS a deep graph generative model that learns a distribution of well-performing architectures.
We show that our proposed graph generator consistently outperforms RNN-based one and achieves better or comparable performances than state-of-the-art NAS methods.
arXiv Detail & Related papers (2022-11-28T09:09:06Z) - NAS-Bench-Graph: Benchmarking Graph Neural Architecture Search [55.75621026447599]
We propose NAS-Bench-Graph, a tailored benchmark that supports unified, reproducible, and efficient evaluations for GraphNAS.
Specifically, we construct a unified, expressive yet compact search space, covering 26,206 unique graph neural network (GNN) architectures.
Based on our proposed benchmark, the performance of GNN architectures can be directly obtained by a look-up table without any further computation.
arXiv Detail & Related papers (2022-06-18T10:17:15Z) - Search to aggregate neighborhood for graph neural network [47.47628113034479]
We propose a framework, which tries to Search to Aggregate NEighborhood (SANE) to automatically design data-specific GNN architectures.
By designing a novel and expressive search space, we propose a differentiable search algorithm, which is more efficient than previous reinforcement learning based methods.
arXiv Detail & Related papers (2021-04-14T03:15:19Z) - Weak NAS Predictors Are All You Need [91.11570424233709]
Recent predictor-based NAS approaches attempt to solve the problem with two key steps: sampling some architecture-performance pairs and fitting a proxy accuracy predictor.
We shift the paradigm from finding a complicated predictor that covers the whole architecture space to a set of weaker predictors that progressively move towards the high-performance sub-space.
Our method costs fewer samples to find the top-performance architectures on NAS-Bench-101 and NAS-Bench-201, and it achieves the state-of-the-art ImageNet performance on the NASNet search space.
arXiv Detail & Related papers (2021-02-21T01:58:43Z) - Neural Architecture Performance Prediction Using Graph Neural Networks [17.224223176258334]
We propose a surrogate model for neural architecture performance prediction built upon Graph Neural Networks (GNN)
We demonstrate the effectiveness of this surrogate model on neural architecture performance prediction for structurally unknown architectures.
arXiv Detail & Related papers (2020-10-19T09:33:57Z) - A Semi-Supervised Assessor of Neural Architectures [157.76189339451565]
We employ an auto-encoder to discover meaningful representations of neural architectures.
A graph convolutional neural network is introduced to predict the performance of architectures.
arXiv Detail & Related papers (2020-05-14T09:02:33Z) - Stage-Wise Neural Architecture Search [65.03109178056937]
Modern convolutional networks such as ResNet and NASNet have achieved state-of-the-art results in many computer vision applications.
These networks consist of stages, which are sets of layers that operate on representations in the same resolution.
It has been demonstrated that increasing the number of layers in each stage improves the prediction ability of the network.
However, the resulting architecture becomes computationally expensive in terms of floating point operations, memory requirements and inference time.
arXiv Detail & Related papers (2020-04-23T14:16:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.