MixPath: A Unified Approach for One-shot Neural Architecture Search
- URL: http://arxiv.org/abs/2001.05887v4
- Date: Wed, 19 Jul 2023 12:58:18 GMT
- Title: MixPath: A Unified Approach for One-shot Neural Architecture Search
- Authors: Xiangxiang Chu, Shun Lu, Xudong Li, Bo Zhang
- Abstract summary: We propose a novel mechanism called Shadow Batch Normalization (SBN) to regularize the disparate feature statistics.
We call our unified multi-path one-shot approach as MixPath, which generates a series of models that achieve state-of-the-art results on ImageNet.
- Score: 13.223963114415552
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Blending multiple convolutional kernels is proved advantageous in neural
architecture design. However, current two-stage neural architecture search
methods are mainly limited to single-path search spaces. How to efficiently
search models of multi-path structures remains a difficult problem. In this
paper, we are motivated to train a one-shot multi-path supernet to accurately
evaluate the candidate architectures. Specifically, we discover that in the
studied search spaces, feature vectors summed from multiple paths are nearly
multiples of those from a single path. Such disparity perturbs the supernet
training and its ranking ability. Therefore, we propose a novel mechanism
called Shadow Batch Normalization (SBN) to regularize the disparate feature
statistics. Extensive experiments prove that SBNs are capable of stabilizing
the optimization and improving ranking performance. We call our unified
multi-path one-shot approach as MixPath, which generates a series of models
that achieve state-of-the-art results on ImageNet.
Related papers
- Real-Time Image Segmentation via Hybrid Convolutional-Transformer Architecture Search [49.81353382211113]
We address the challenge of integrating multi-head self-attention into high resolution representation CNNs efficiently.
We develop a multi-target multi-branch supernet method, which fully utilizes the advantages of high-resolution features.
We present a series of model via Hybrid Convolutional-Transformer Architecture Search (HyCTAS) method that searched for the best hybrid combination of light-weight convolution layers and memory-efficient self-attention layers.
arXiv Detail & Related papers (2024-03-15T15:47:54Z) - OFA$^2$: A Multi-Objective Perspective for the Once-for-All Neural
Architecture Search [79.36688444492405]
Once-for-All (OFA) is a Neural Architecture Search (NAS) framework designed to address the problem of searching efficient architectures for devices with different resources constraints.
We aim to give one step further in the search for efficiency by explicitly conceiving the search stage as a multi-objective optimization problem.
arXiv Detail & Related papers (2023-03-23T21:30:29Z) - Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - Mutually-aware Sub-Graphs Differentiable Architecture Search [29.217547815683748]
Mutually-aware Sub-Graphs Differentiable Architecture Search (MSG-DAS)
MSG-DAS is a differentiable Gumbel-TopK sampler that produces multiple mutually exclusive single-path sub-graphs.
We demonstrate the effectiveness of our methods on ImageNet and CIFAR10, where the searched models show a comparable performance as the most recent approaches.
arXiv Detail & Related papers (2021-07-09T09:31:31Z) - Cream of the Crop: Distilling Prioritized Paths For One-Shot Neural
Architecture Search [60.965024145243596]
One-shot weight sharing methods have recently drawn great attention in neural architecture search due to high efficiency and competitive performance.
To alleviate this problem, we present a simple yet effective architecture distillation method.
We introduce the concept of prioritized path, which refers to the architecture candidates exhibiting superior performance during training.
Since the prioritized paths are changed on the fly depending on their performance and complexity, the final obtained paths are the cream of the crop.
arXiv Detail & Related papers (2020-10-29T17:55:05Z) - Neural Architecture Search as Sparse Supernet [78.09905626281046]
This paper aims at enlarging the problem of Neural Architecture Search (NAS) from Single-Path and Multi-Path Search to automated Mixed-Path Search.
We model the NAS problem as a sparse supernet using a new continuous architecture representation with a mixture of sparsity constraints.
The sparse supernet enables us to automatically achieve sparsely-mixed paths upon a compact set of nodes.
arXiv Detail & Related papers (2020-07-31T14:51:52Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.