Bag of Baselines for Multi-objective Joint Neural Architecture Search
and Hyperparameter Optimization
- URL: http://arxiv.org/abs/2105.01015v1
- Date: Mon, 3 May 2021 17:04:56 GMT
- Title: Bag of Baselines for Multi-objective Joint Neural Architecture Search
and Hyperparameter Optimization
- Authors: Julia Guerrero-Viu, Sven Hauns, Sergio Izquierdo, Guilherme Miotto,
Simon Schrodi, Andre Biedenkapp, Thomas Elsken, Difan Deng, Marius Lindauer,
Frank Hutter
- Abstract summary: Neural architecture search (NAS) and hyper parameter optimization (HPO) make deep learning accessible to non-experts.
We propose a set of methods that extend current approaches to jointly optimize neural architectures and hyper parameters with respect to multiple objectives.
These methods will serve as simple baselines for future research on multi-objective joint NAS + HPO.
- Score: 29.80410614305851
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural architecture search (NAS) and hyperparameter optimization (HPO) make
deep learning accessible to non-experts by automatically finding the
architecture of the deep neural network to use and tuning the hyperparameters
of the used training pipeline. While both NAS and HPO have been studied
extensively in recent years, NAS methods typically assume fixed hyperparameters
and vice versa - there exists little work on joint NAS + HPO. Furthermore, NAS
has recently often been framed as a multi-objective optimization problem, in
order to take, e.g., resource requirements into account. In this paper, we
propose a set of methods that extend current approaches to jointly optimize
neural architectures and hyperparameters with respect to multiple objectives.
We hope that these methods will serve as simple baselines for future research
on multi-objective joint NAS + HPO. To facilitate this, all our code is
available at https://github.com/automl/multi-obj-baselines.
Related papers
- Efficient Multi-Objective Neural Architecture Search via Pareto Dominance-based Novelty Search [0.0]
Neural Architecture Search (NAS) aims to automate the discovery of high-performing deep neural network architectures.
Traditional NAS approaches typically optimize a certain performance metric (e.g., prediction accuracy) overlooking large parts of the architecture search space that potentially contain interesting network configurations.
This paper presents a novelty search for multi-objective NAS with Multiple Training-Free metrics (MTF-PDNS)
arXiv Detail & Related papers (2024-07-30T08:52:10Z) - DNA Family: Boosting Weight-Sharing NAS with Block-Wise Supervisions [121.05720140641189]
We develop a family of models with the distilling neural architecture (DNA) techniques.
Our proposed DNA models can rate all architecture candidates, as opposed to previous works that can only access a sub- search space using algorithms.
Our models achieve state-of-the-art top-1 accuracy of 78.9% and 83.6% on ImageNet for a mobile convolutional network and a small vision transformer, respectively.
arXiv Detail & Related papers (2024-03-02T22:16:47Z) - A Survey on Multi-Objective Neural Architecture Search [9.176056742068813]
Multi-Objective Neural architecture Search (MONAS) has been attracting attentions.
We present an overview of principal and state-of-the-art works in the field of MONAS.
arXiv Detail & Related papers (2023-07-18T09:42:51Z) - DiffusionNAG: Predictor-guided Neural Architecture Generation with Diffusion Models [56.584561770857306]
We propose a novel conditional Neural Architecture Generation (NAG) framework based on diffusion models, dubbed DiffusionNAG.
Specifically, we consider the neural architectures as directed graphs and propose a graph diffusion model for generating them.
We validate the effectiveness of DiffusionNAG through extensive experiments in two predictor-based NAS scenarios: Transferable NAS and Bayesian Optimization (BO)-based NAS.
When integrated into a BO-based algorithm, DiffusionNAG outperforms existing BO-based NAS approaches, particularly in the large MobileNetV3 search space on the ImageNet 1K dataset.
arXiv Detail & Related papers (2023-05-26T13:58:18Z) - Generalization Properties of NAS under Activation and Skip Connection
Search [66.8386847112332]
We study the generalization properties of Neural Architecture Search (NAS) under a unifying framework.
We derive the lower (and upper) bounds of the minimum eigenvalue of the Neural Tangent Kernel (NTK) under the (in)finite-width regime.
We show how the derived results can guide NAS to select the top-performing architectures, even in the case without training.
arXiv Detail & Related papers (2022-09-15T12:11:41Z) - HyperSegNAS: Bridging One-Shot Neural Architecture Search with 3D
Medical Image Segmentation using HyperNet [51.60655410423093]
We introduce HyperSegNAS to enable one-shot Neural Architecture Search (NAS) for medical image segmentation.
We show that HyperSegNAS yields better performing and more intuitive architectures compared to the previous state-of-the-art (SOTA) segmentation networks.
Our method is evaluated on public datasets from the Medical Decathlon (MSD) challenge, and achieves SOTA performances.
arXiv Detail & Related papers (2021-12-20T16:21:09Z) - TND-NAS: Towards Non-differentiable Objectives in Progressive
Differentiable NAS Framework [6.895590095853327]
Differentiable architecture search has gradually become the mainstream research topic in the field of Neural Architecture Search (NAS)
Recent differentiable NAS also aims at further improving the search performance and reducing the GPU-memory consumption.
We propose the TND-NAS, which is with the merits of the high efficiency in differentiable NAS framework and the compatibility among non-differentiable metrics in Multi-objective NAS.
arXiv Detail & Related papers (2021-11-06T14:19:36Z) - NAS-HPO-Bench-II: A Benchmark Dataset on Joint Optimization of
Convolutional Neural Network Architecture and Training Hyperparameters [4.039245878626346]
This paper introduces the first benchmark dataset for joint optimization of network connections and training hyperparameters, which we call NAS-HPO-Bench-II.
We collect the performance data of 4K cell-based convolutional neural network architectures trained on the CIFAR-10 dataset with different learning rate and batch size settings.
We build a surrogate model predicting the accuracies after 200 epoch training to provide the performance data of longer training epoch.
arXiv Detail & Related papers (2021-10-19T18:00:01Z) - DHA: End-to-End Joint Optimization of Data Augmentation Policy,
Hyper-parameter and Architecture [81.82173855071312]
We propose an end-to-end solution that integrates the AutoML components and returns a ready-to-use model at the end of the search.
Dha achieves state-of-the-art (SOTA) results on various datasets, especially 77.4% accuracy on ImageNet with cell based search space.
arXiv Detail & Related papers (2021-09-13T08:12:50Z) - Hyperparameter Optimization in Neural Networks via Structured Sparse
Recovery [54.60327265077322]
We study two important problems in the automated design of neural networks through the lens of sparse recovery methods.
In the first part of this paper, we establish a novel connection between HPO and structured sparse recovery.
In the second part of this paper, we establish a connection between NAS and structured sparse recovery.
arXiv Detail & Related papers (2020-07-07T00:57:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.