Warm-starting DARTS using meta-learning
- URL: http://arxiv.org/abs/2205.06355v1
- Date: Thu, 12 May 2022 20:40:26 GMT
- Title: Warm-starting DARTS using meta-learning
- Authors: Matej Grobelnik and Joaquin Vanschoren
- Abstract summary: Neural architecture search (NAS) has shown great promise in the field of automated machine learning (AutoML)
We present a meta-learning framework to warm-start Differentiable architecture search (DARTS)
- Score: 4.035753155957698
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Neural architecture search (NAS) has shown great promise in the field of
automated machine learning (AutoML). NAS has outperformed hand-designed
networks and made a significant step forward in the field of automating the
design of deep neural networks, thus further reducing the need for human
expertise. However, most research is done targeting a single specific task,
leaving research of NAS methods over multiple tasks mostly overlooked.
Generally, there exist two popular ways to find an architecture for some novel
task. Either searching from scratch, which is ineffective by design, or
transferring discovered architectures from other tasks, which provides no
performance guarantees and is probably not optimal. In this work, we present a
meta-learning framework to warm-start Differentiable architecture search
(DARTS). DARTS is a NAS method that can be initialized with a transferred
architecture and is able to quickly adapt to new tasks. A task similarity
measure is used to determine which transfer architecture is selected, as
transfer architectures found on similar tasks will likely perform better.
Additionally, we employ a simple meta-transfer architecture that was learned
over multiple tasks. Experiments show that warm-started DARTS is able to find
competitive performing architectures while reducing searching costs on average
by 60%.
Related papers
- EM-DARTS: Hierarchical Differentiable Architecture Search for Eye Movement Recognition [54.99121380536659]
Eye movement biometrics have received increasing attention thanks to its high secure identification.
Deep learning (DL) models have been recently successfully applied for eye movement recognition.
DL architecture still is determined by human prior knowledge.
We propose EM-DARTS, a hierarchical differentiable architecture search algorithm to automatically design the DL architecture for eye movement recognition.
arXiv Detail & Related papers (2024-09-22T13:11:08Z) - Design Principle Transfer in Neural Architecture Search via Large Language Models [37.004026595537006]
Transferable neural architecture search (TNAS) has been introduced to design efficient neural architectures for multiple tasks.
In TNAS, architectural knowledge accumulated in previous search processes is reused to warm up the architecture search for new tasks.
This work proposes a novel transfer paradigm, i.e., design principle transfer.
arXiv Detail & Related papers (2024-08-21T04:27:44Z) - A Pairwise Comparison Relation-assisted Multi-objective Evolutionary Neural Architecture Search Method with Multi-population Mechanism [58.855741970337675]
Neural architecture search (NAS) enables re-searchers to automatically explore vast search spaces and find efficient neural networks.
NAS suffers from a key bottleneck, i.e., numerous architectures need to be evaluated during the search process.
We propose the SMEM-NAS, a pairwise com-parison relation-assisted multi-objective evolutionary algorithm based on a multi-population mechanism.
arXiv Detail & Related papers (2024-07-22T12:46:22Z) - Across-Task Neural Architecture Search via Meta Learning [1.225795556154044]
Adequate labeled data and expensive compute resources are the prerequisites for the success of neural architecture search(NAS)
It is challenging to apply NAS in meta-learning scenarios with limited compute resources and data.
In this paper, an across-task neural architecture search (AT-NAS) is proposed to address the problem through combining gradient-based meta-learning with EA-based NAS.
arXiv Detail & Related papers (2021-10-12T09:07:33Z) - Elastic Architecture Search for Diverse Tasks with Different Resources [87.23061200971912]
We study a new challenging problem of efficient deployment for diverse tasks with different resources, where the resource constraint and task of interest corresponding to a group of classes are dynamically specified at testing time.
Previous NAS approaches seek to design architectures for all classes simultaneously, which may not be optimal for some individual tasks.
We present a novel and general framework, called Elastic Architecture Search (EAS), permitting instant specializations at runtime for diverse tasks with various resource constraints.
arXiv Detail & Related papers (2021-08-03T00:54:27Z) - Rapid Neural Architecture Search by Learning to Generate Graphs from
Datasets [42.993720854755736]
We propose an efficient Neural Search (NAS) framework that is trained once on a database consisting of datasets and pretrained networks.
We show that our model meta-learned on subsets of ImageNet-1K and architectures from NAS-Bench 201 search space successfully generalizes to multiple unseen datasets.
arXiv Detail & Related papers (2021-07-02T06:33:59Z) - Stage-Wise Neural Architecture Search [65.03109178056937]
Modern convolutional networks such as ResNet and NASNet have achieved state-of-the-art results in many computer vision applications.
These networks consist of stages, which are sets of layers that operate on representations in the same resolution.
It has been demonstrated that increasing the number of layers in each stage improves the prediction ability of the network.
However, the resulting architecture becomes computationally expensive in terms of floating point operations, memory requirements and inference time.
arXiv Detail & Related papers (2020-04-23T14:16:39Z) - MTL-NAS: Task-Agnostic Neural Architecture Search towards
General-Purpose Multi-Task Learning [71.90902837008278]
We propose to incorporate neural architecture search (NAS) into general-purpose multi-task learning (GP-MTL)
In order to adapt to different task combinations, we disentangle the GP-MTL networks into single-task backbones.
We also propose a novel single-shot gradient-based search algorithm that closes the performance gap between the searched architectures.
arXiv Detail & Related papers (2020-03-31T09:49:14Z) - DA-NAS: Data Adapted Pruning for Efficient Neural Architecture Search [76.9225014200746]
Efficient search is a core issue in Neural Architecture Search (NAS)
We present DA-NAS that can directly search the architecture for large-scale target tasks while allowing a large candidate set in a more efficient manner.
It is 2x faster than previous methods while the accuracy is currently state-of-the-art, at 76.2% under small FLOPs constraint.
arXiv Detail & Related papers (2020-03-27T17:55:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.