FBNetV5: Neural Architecture Search for Multiple Tasks in One Run
- URL: http://arxiv.org/abs/2111.10007v1
- Date: Fri, 19 Nov 2021 02:07:34 GMT
- Title: FBNetV5: Neural Architecture Search for Multiple Tasks in One Run
- Authors: Bichen Wu, Chaojian Li, Hang Zhang, Xiaoliang Dai, Peizhao Zhang,
Matthew Yu, Jialiang Wang, Yingyan Lin, Peter Vajda
- Abstract summary: We propose FBNetV5, a framework that can search for neural architectures for a variety of vision tasks with much reduced computational cost and human effort.
Specifically, we design 1) a search space that is simple yet inclusive and transferable; 2) a multitask search process that is disentangled with target tasks' training pipeline; and 3) an algorithm to simultaneously search for architectures for multiple tasks with a computational cost agnostic to the number of tasks.
We evaluate the proposed FBNetV5 targeting three fundamental vision tasks -- image classification, object detection, and semantic segmentation.
- Score: 28.645664534127516
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural Architecture Search (NAS) has been widely adopted to design accurate
and efficient image classification models. However, applying NAS to a new
computer vision task still requires a huge amount of effort. This is because 1)
previous NAS research has been over-prioritized on image classification while
largely ignoring other tasks; 2) many NAS works focus on optimizing
task-specific components that cannot be favorably transferred to other tasks;
and 3) existing NAS methods are typically designed to be "proxyless" and
require significant effort to be integrated with each new task's training
pipelines. To tackle these challenges, we propose FBNetV5, a NAS framework that
can search for neural architectures for a variety of vision tasks with much
reduced computational cost and human effort. Specifically, we design 1) a
search space that is simple yet inclusive and transferable; 2) a multitask
search process that is disentangled with target tasks' training pipeline; and
3) an algorithm to simultaneously search for architectures for multiple tasks
with a computational cost agnostic to the number of tasks. We evaluate the
proposed FBNetV5 targeting three fundamental vision tasks -- image
classification, object detection, and semantic segmentation. Models searched by
FBNetV5 in a single run of search have outperformed the previous
stateof-the-art in all the three tasks: image classification (e.g., +1.3%
ImageNet top-1 accuracy under the same FLOPs as compared to FBNetV3), semantic
segmentation (e.g., +1.8% higher ADE20K val. mIoU than SegFormer with 3.6x
fewer FLOPs), and object detection (e.g., +1.1% COCO val. mAP with 1.2x fewer
FLOPs as compared to YOLOX).
Related papers
- Arch-Graph: Acyclic Architecture Relation Predictor for
Task-Transferable Neural Architecture Search [96.31315520244605]
Arch-Graph is a transferable NAS method that predicts task-specific optimal architectures.
We show Arch-Graph's transferability and high sample efficiency across numerous tasks.
It is able to find top 0.16% and 0.29% architectures on average on two search spaces under the budget of only 50 models.
arXiv Detail & Related papers (2022-04-12T16:46:06Z) - NAS-Bench-360: Benchmarking Diverse Tasks for Neural Architecture Search [18.9676056830197]
Most existing neural architecture search (NAS) benchmarks and algorithms prioritize performance on well-studied tasks.
We present NAS-Bench-360, a benchmark suite for evaluating state-of-the-art NAS methods for convolutional neural networks (CNNs)
arXiv Detail & Related papers (2021-10-12T01:13:18Z) - TransNAS-Bench-101: Improving Transferability and Generalizability of
Cross-Task Neural Architecture Search [98.22779489340869]
We propose TransNAS-Bench-101, a benchmark dataset containing network performance across seven vision tasks.
We explore two fundamentally different types of search space: cell-level search space and macro-level search space.
With 7,352 backbones evaluated on seven tasks, 51,464 trained models with detailed training information are provided.
arXiv Detail & Related papers (2021-05-25T12:15:21Z) - Learning Versatile Neural Architectures by Propagating Network Codes [74.2450894473073]
We propose a novel "neural predictor", which is able to predict an architecture's performance in multiple datasets and tasks.
NCP learns from network codes but not original data, enabling it to update the architecture efficiently across datasets.
arXiv Detail & Related papers (2021-03-24T15:20:38Z) - Neural Architecture Search on ImageNet in Four GPU Hours: A
Theoretically Inspired Perspective [88.39981851247727]
We propose a novel framework called training-free neural architecture search (TE-NAS)
TE-NAS ranks architectures by analyzing the spectrum of the neural tangent kernel (NTK) and the number of linear regions in the input space.
We show that: (1) these two measurements imply the trainability and expressivity of a neural network; (2) they strongly correlate with the network's test accuracy.
arXiv Detail & Related papers (2021-02-23T07:50:44Z) - FNA++: Fast Network Adaptation via Parameter Remapping and Architecture
Search [35.61441231491448]
We propose a Fast Network Adaptation (FNA++) method, which can adapt both the architecture and parameters of a seed network.
In our experiments, we apply FNA++ on MobileNetV2 to obtain new networks for semantic segmentation, object detection, and human pose estimation.
The total computation cost of FNA++ is significantly less than SOTA segmentation and detection NAS approaches.
arXiv Detail & Related papers (2020-06-21T10:03:34Z) - MTL-NAS: Task-Agnostic Neural Architecture Search towards
General-Purpose Multi-Task Learning [71.90902837008278]
We propose to incorporate neural architecture search (NAS) into general-purpose multi-task learning (GP-MTL)
In order to adapt to different task combinations, we disentangle the GP-MTL networks into single-task backbones.
We also propose a novel single-shot gradient-based search algorithm that closes the performance gap between the searched architectures.
arXiv Detail & Related papers (2020-03-31T09:49:14Z) - Fast Neural Network Adaptation via Parameter Remapping and Architecture
Search [35.61441231491448]
Deep neural networks achieve remarkable performance in many computer vision tasks.
Most state-of-the-art (SOTA) semantic segmentation and object detection approaches reuse neural network architectures designed for image classification as the backbone.
One major challenge though, is that ImageNet pre-training of the search space representation incurs huge computational cost.
In this paper, we propose a Fast Neural Network Adaptation (FNA) method, which can adapt both the architecture and parameters of a seed network.
arXiv Detail & Related papers (2020-01-08T13:45:15Z) - Scalable NAS with Factorizable Architectural Parameters [102.51428615447703]
Neural Architecture Search (NAS) is an emerging topic in machine learning and computer vision.
This paper presents a scalable algorithm by factorizing a large set of candidate operators into smaller subspaces.
With a small increase in search costs and no extra costs in re-training, we find interesting architectures that were not explored before.
arXiv Detail & Related papers (2019-12-31T10:26:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.