OFA$^2$: A Multi-Objective Perspective for the Once-for-All Neural
Architecture Search
- URL: http://arxiv.org/abs/2303.13683v1
- Date: Thu, 23 Mar 2023 21:30:29 GMT
- Title: OFA$^2$: A Multi-Objective Perspective for the Once-for-All Neural
Architecture Search
- Authors: Rafael C. Ito and Fernando J. Von Zuben
- Abstract summary: Once-for-All (OFA) is a Neural Architecture Search (NAS) framework designed to address the problem of searching efficient architectures for devices with different resources constraints.
We aim to give one step further in the search for efficiency by explicitly conceiving the search stage as a multi-objective optimization problem.
- Score: 79.36688444492405
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Once-for-All (OFA) is a Neural Architecture Search (NAS) framework designed
to address the problem of searching efficient architectures for devices with
different resources constraints by decoupling the training and the searching
stages. The computationally expensive process of training the OFA neural
network is done only once, and then it is possible to perform multiple searches
for subnetworks extracted from this trained network according to each
deployment scenario. In this work we aim to give one step further in the search
for efficiency by explicitly conceiving the search stage as a multi-objective
optimization problem. A Pareto frontier is then populated with efficient, and
already trained, neural architectures exhibiting distinct trade-offs among the
conflicting objectives. This could be achieved by using any multi-objective
evolutionary algorithm during the search stage, such as NSGA-II and SMS-EMOA.
In other words, the neural network is trained once, the searching for
subnetworks considering different hardware constraints is also done one single
time, and then the user can choose a suitable neural network according to each
deployment scenario. The conjugation of OFA and an explicit algorithm for
multi-objective optimization opens the possibility of a posteriori
decision-making in NAS, after sampling efficient subnetworks which are a very
good approximation of the Pareto frontier, given that those subnetworks are
already trained and ready to use. The source code and the final search
algorithm will be released at https://github.com/ito-rafael/once-for-all-2
Related papers
- SimQ-NAS: Simultaneous Quantization Policy and Neural Architecture
Search [6.121126813817338]
Recent one-shot Neural Architecture Search algorithms rely on training a hardware-agnostic super-network tailored to a specific task and then extracting efficient sub-networks for different hardware platforms.
We show that by using multi-objective search algorithms paired with lightly trained predictors, we can efficiently search for both the sub-network architecture and the corresponding quantization policy.
arXiv Detail & Related papers (2023-12-19T22:08:49Z) - Elastic Architecture Search for Diverse Tasks with Different Resources [87.23061200971912]
We study a new challenging problem of efficient deployment for diverse tasks with different resources, where the resource constraint and task of interest corresponding to a group of classes are dynamically specified at testing time.
Previous NAS approaches seek to design architectures for all classes simultaneously, which may not be optimal for some individual tasks.
We present a novel and general framework, called Elastic Architecture Search (EAS), permitting instant specializations at runtime for diverse tasks with various resource constraints.
arXiv Detail & Related papers (2021-08-03T00:54:27Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z) - ENAS4D: Efficient Multi-stage CNN Architecture Search for Dynamic
Inference [12.15628508314535]
We introduce a general framework, ENAS4D, which can efficiently search for optimal multi-stage CNN architecture.
Experiments on the ImageNet classification task demonstrate that the multi-stage CNNs searched by ENAS4D consistently outperform the state-of-the-art method for dyanmic inference.
arXiv Detail & Related papers (2020-09-19T08:08:26Z) - CATCH: Context-based Meta Reinforcement Learning for Transferrable
Architecture Search [102.67142711824748]
CATCH is a novel Context-bAsed meTa reinforcement learning algorithm for transferrable arChitecture searcH.
The combination of meta-learning and RL allows CATCH to efficiently adapt to new tasks while being agnostic to search spaces.
It is also capable of handling cross-domain architecture search as competitive networks on ImageNet, COCO, and Cityscapes are identified.
arXiv Detail & Related papers (2020-07-18T09:35:53Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - MTL-NAS: Task-Agnostic Neural Architecture Search towards
General-Purpose Multi-Task Learning [71.90902837008278]
We propose to incorporate neural architecture search (NAS) into general-purpose multi-task learning (GP-MTL)
In order to adapt to different task combinations, we disentangle the GP-MTL networks into single-task backbones.
We also propose a novel single-shot gradient-based search algorithm that closes the performance gap between the searched architectures.
arXiv Detail & Related papers (2020-03-31T09:49:14Z) - PONAS: Progressive One-shot Neural Architecture Search for Very
Efficient Deployment [9.442139459221783]
We propose Progressive One-shot Neural Architecture Search (PONAS) that combines advantages of progressive NAS and one-shot methods.
PONAS is able to find architecture of a specialized network in around 10 seconds.
In ImageNet classification, 75.2% top-1 accuracy can be obtained, which is comparable with the state of the arts.
arXiv Detail & Related papers (2020-03-11T05:00:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.