Landmark Regularization: Ranking Guided Super-Net Training in Neural
Architecture Search
- URL: http://arxiv.org/abs/2104.05309v1
- Date: Mon, 12 Apr 2021 09:32:33 GMT
- Title: Landmark Regularization: Ranking Guided Super-Net Training in Neural
Architecture Search
- Authors: Kaicheng Yu, Rene Ranftl, Mathieu Salzmann
- Abstract summary: Weight sharing has become a de facto standard in neural architecture search because it enables the search to be done on commodity hardware.
Recent works have empirically shown a ranking disorder between the performance of stand-alone architectures and that of the corresponding shared-weight networks.
We propose a regularization term that aims to maximize the correlation between the performance rankings of the shared-weight network and that of the standalone architectures.
- Score: 70.57382341642418
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Weight sharing has become a de facto standard in neural architecture search
because it enables the search to be done on commodity hardware. However, recent
works have empirically shown a ranking disorder between the performance of
stand-alone architectures and that of the corresponding shared-weight networks.
This violates the main assumption of weight-sharing NAS algorithms, thus
limiting their effectiveness. We tackle this issue by proposing a
regularization term that aims to maximize the correlation between the
performance rankings of the shared-weight network and that of the standalone
architectures using a small set of landmark architectures. We incorporate our
regularization term into three different NAS algorithms and show that it
consistently improves performance across algorithms, search-spaces, and tasks.
Related papers
- Efficient Multi-Objective Neural Architecture Search via Pareto Dominance-based Novelty Search [0.0]
Neural Architecture Search (NAS) aims to automate the discovery of high-performing deep neural network architectures.
Traditional NAS approaches typically optimize a certain performance metric (e.g., prediction accuracy) overlooking large parts of the architecture search space that potentially contain interesting network configurations.
This paper presents a novelty search for multi-objective NAS with Multiple Training-Free metrics (MTF-PDNS)
arXiv Detail & Related papers (2024-07-30T08:52:10Z) - A General-Purpose Transferable Predictor for Neural Architecture Search [22.883809911265445]
We propose a general-purpose neural predictor for Neural Architecture Search (NAS) that can transfer across search spaces.
Experimental results on NAS-Bench-101, 201 and 301 demonstrate the efficacy of our scheme.
arXiv Detail & Related papers (2023-02-21T17:28:05Z) - Generalizing Few-Shot NAS with Gradient Matching [165.5690495295074]
One-Shot methods train one supernet to approximate the performance of every architecture in the search space via weight-sharing.
Few-Shot NAS reduces the level of weight-sharing by splitting the One-Shot supernet into multiple separated sub-supernets.
It significantly outperforms its Few-Shot counterparts while surpassing previous comparable methods in terms of the accuracy of derived architectures.
arXiv Detail & Related papers (2022-03-29T03:06:16Z) - An Analysis of Super-Net Heuristics in Weight-Sharing NAS [70.57382341642418]
We show that simple random search achieves competitive performance to complex state-of-the-art NAS algorithms when the super-net is properly trained.
We show that simple random search achieves competitive performance to complex state-of-the-art NAS algorithms when the super-net is properly trained.
arXiv Detail & Related papers (2021-10-04T02:18:44Z) - On Constrained Optimization in Differentiable Neural Architecture Search [3.0682439731292592]
Differentiable Architecture Search (DARTS) is a recently proposed neural architecture search (NAS) method based on a differentiable relaxation.
We propose and analyze three improvements to architectural weight competition, update scheduling, and regularization towards discretization.
arXiv Detail & Related papers (2021-06-22T10:22:31Z) - Hierarchical Neural Architecture Search for Deep Stereo Matching [131.94481111956853]
We propose the first end-to-end hierarchical NAS framework for deep stereo matching.
Our framework incorporates task-specific human knowledge into the neural architecture search framework.
It is ranked at the top 1 accuracy on KITTI stereo 2012, 2015 and Middlebury benchmarks, as well as the top 1 on SceneFlow dataset.
arXiv Detail & Related papers (2020-10-26T11:57:37Z) - Fitting the Search Space of Weight-sharing NAS with Graph Convolutional
Networks [100.14670789581811]
We train a graph convolutional network to fit the performance of sampled sub-networks.
With this strategy, we achieve a higher rank correlation coefficient in the selected set of candidates.
arXiv Detail & Related papers (2020-04-17T19:12:39Z) - Scalable NAS with Factorizable Architectural Parameters [102.51428615447703]
Neural Architecture Search (NAS) is an emerging topic in machine learning and computer vision.
This paper presents a scalable algorithm by factorizing a large set of candidate operators into smaller subspaces.
With a small increase in search costs and no extra costs in re-training, we find interesting architectures that were not explored before.
arXiv Detail & Related papers (2019-12-31T10:26:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.