Multi-Predict: Few Shot Predictors For Efficient Neural Architecture
Search
- URL: http://arxiv.org/abs/2306.02459v1
- Date: Sun, 4 Jun 2023 20:22:14 GMT
- Title: Multi-Predict: Few Shot Predictors For Efficient Neural Architecture
Search
- Authors: Yash Akhauri, Mohamed S. Abdelfattah
- Abstract summary: We introduce a novel search-space independent NN encoding based on zero-cost proxies that achieves sample-efficient prediction on multiple tasks and NAS search spaces.
Our NN encoding enables multi-search-space transfer of latency predictors from NASBench-201 to FBNet in under 85 HW measurements.
- Score: 10.538869116366415
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Many hardware-aware neural architecture search (NAS) methods have been
developed to optimize the topology of neural networks (NN) with the joint
objectives of higher accuracy and lower latency. Recently, both accuracy and
latency predictors have been used in NAS with great success, achieving high
sample efficiency and accurate modeling of hardware (HW) device latency
respectively. However, a new accuracy predictor needs to be trained for every
new NAS search space or NN task, and a new latency predictor needs to be
additionally trained for every new HW device. In this paper, we explore methods
to enable multi-task, multi-search-space, and multi-HW adaptation of accuracy
and latency predictors to reduce the cost of NAS. We introduce a novel
search-space independent NN encoding based on zero-cost proxies that achieves
sample-efficient prediction on multiple tasks and NAS search spaces, improving
the end-to-end sample efficiency of latency and accuracy predictors by over an
order of magnitude in multiple scenarios. For example, our NN encoding enables
multi-search-space transfer of latency predictors from NASBench-201 to FBNet
(and vice-versa) in under 85 HW measurements, a 400$\times$ improvement in
sample efficiency compared to a recent meta-learning approach. Our method also
improves the total sample efficiency of accuracy predictors by over an order of
magnitude. Finally, we demonstrate the effectiveness of our method for
multi-search-space and multi-task accuracy prediction on 28 NAS search spaces
and tasks.
Related papers
- A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - On Latency Predictors for Neural Architecture Search [8.564763702766776]
We introduce a comprehensive suite of latency prediction tasks obtained in a principled way through automated partitioning of hardware device sets.
We then design a general latency predictor to comprehensively study (1) the predictor architecture, (2) NN sample selection methods, (3) hardware device representations, and (4) NN operation encoding schemes.
Building on conclusions from our study, we present an end-to-end latency predictor training strategy.
arXiv Detail & Related papers (2024-03-04T19:59:32Z) - IQNAS: Interpretable Integer Quadratic Programming Neural Architecture
Search [40.77061519007659]
A popular approach to find fitting networks is through constrained Neural Architecture Search (NAS)
Previous methods use complicated predictors for the accuracy of the network.
We introduce Interpretable Quadratic programming Neural Architecture Search (IQNAS)
arXiv Detail & Related papers (2021-10-24T09:45:00Z) - Efficient Sampling for Predictor-Based Neural Architecture Search [3.287802528135173]
We study predictor-based NAS algorithms for neural architecture search.
We show that the sample efficiency of predictor-based algorithms decreases dramatically if the proxy is only computed for a subset of the search space.
This is an important step to make predictor-based NAS algorithms useful, in practice.
arXiv Detail & Related papers (2020-11-24T11:36:36Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - BRP-NAS: Prediction-based NAS using GCNs [21.765796576990137]
BRP-NAS is an efficient hardware-aware NAS enabled by an accurate performance predictor-based on graph convolutional network (GCN)
We show that our proposed method outperforms all prior methods on NAS-Bench-101 and NAS-Bench-201.
We also release LatBench -- a latency dataset of NAS-Bench-201 models running on a broad range of devices.
arXiv Detail & Related papers (2020-07-16T21:58:43Z) - Accuracy Prediction with Non-neural Model for Neural Architecture Search [185.0651567642238]
We study an alternative approach which uses non-neural model for accuracy prediction.
We leverage gradient boosting decision tree (GBDT) as the predictor for Neural architecture search (NAS)
Experiments on NASBench-101 and ImageNet demonstrate the effectiveness of using GBDT as predictor for NAS.
arXiv Detail & Related papers (2020-07-09T13:28:49Z) - APQ: Joint Search for Network Architecture, Pruning and Quantization
Policy [49.3037538647714]
We present APQ for efficient deep learning inference on resource-constrained hardware.
Unlike previous methods that separately search the neural architecture, pruning policy, and quantization policy, we optimize them in a joint manner.
With the same accuracy, APQ reduces the latency/energy by 2x/1.3x over MobileNetV2+HAQ.
arXiv Detail & Related papers (2020-06-15T16:09:17Z) - FBNetV3: Joint Architecture-Recipe Search using Predictor Pretraining [65.39532971991778]
We present an accuracy predictor that scores architecture and training recipes jointly, guiding both sample selection and ranking.
We run fast evolutionary searches in just CPU minutes to generate architecture-recipe pairs for a variety of resource constraints.
FBNetV3 makes up a family of state-of-the-art compact neural networks that outperform both automatically and manually-designed competitors.
arXiv Detail & Related papers (2020-06-03T05:20:21Z) - Widening and Squeezing: Towards Accurate and Efficient QNNs [125.172220129257]
Quantization neural networks (QNNs) are very attractive to the industry because their extremely cheap calculation and storage overhead, but their performance is still worse than that of networks with full-precision parameters.
Most of existing methods aim to enhance performance of QNNs especially binary neural networks by exploiting more effective training techniques.
We address this problem by projecting features in original full-precision networks to high-dimensional quantization features.
arXiv Detail & Related papers (2020-02-03T04:11:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.