Shapley-NAS: Discovering Operation Contribution for Neural Architecture
Search
- URL: http://arxiv.org/abs/2206.09811v1
- Date: Mon, 20 Jun 2022 14:41:49 GMT
- Title: Shapley-NAS: Discovering Operation Contribution for Neural Architecture
Search
- Authors: Han Xiao, Ziwei Wang, Zheng Zhu, Jie Zhou, Jiwen Lu
- Abstract summary: We propose a Shapley value based method to evaluate operation contribution (Shapley-NAS) for neural architecture search.
We show that our method outperforms the state-of-the-art methods by a considerable margin with light search cost.
- Score: 96.20505710087392
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a Shapley value based method to evaluate operation
contribution (Shapley-NAS) for neural architecture search. Differentiable
architecture search (DARTS) acquires the optimal architectures by optimizing
the architecture parameters with gradient descent, which significantly reduces
the search cost. However, the magnitude of architecture parameters updated by
gradient descent fails to reveal the actual operation importance to the task
performance and therefore harms the effectiveness of obtained architectures. By
contrast, we propose to evaluate the direct influence of operations on
validation accuracy. To deal with the complex relationships between supernet
components, we leverage Shapley value to quantify their marginal contributions
by considering all possible combinations. Specifically, we iteratively optimize
the supernet weights and update the architecture parameters by evaluating
operation contributions via Shapley value, so that the optimal architectures
are derived by selecting the operations that contribute significantly to the
tasks. Since the exact computation of Shapley value is NP-hard, the Monte-Carlo
sampling based algorithm with early truncation is employed for efficient
approximation, and the momentum update mechanism is adopted to alleviate
fluctuation of the sampling process. Extensive experiments on various datasets
and various search spaces show that our Shapley-NAS outperforms the
state-of-the-art methods by a considerable margin with light search cost. The
code is available at https://github.com/Euphoria16/Shapley-NAS.git
Related papers
- Neural Architecture Search using Particle Swarm and Ant Colony
Optimization [0.0]
This paper focuses on training and optimizing CNNs using the Swarm Intelligence (SI) components of OpenNAS.
A system integrating open source tools for Neural Architecture Search (OpenNAS), in the classification of images, has been developed.
arXiv Detail & Related papers (2024-03-06T15:23:26Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - DAAS: Differentiable Architecture and Augmentation Policy Search [107.53318939844422]
This work considers the possible coupling between neural architectures and data augmentation and proposes an effective algorithm jointly searching for them.
Our approach achieves 97.91% accuracy on CIFAR-10 and 76.6% Top-1 accuracy on ImageNet dataset, showing the outstanding performance of our search algorithm.
arXiv Detail & Related papers (2021-09-30T17:15:17Z) - Rethinking Architecture Selection in Differentiable NAS [74.61723678821049]
Differentiable Neural Architecture Search is one of the most popular NAS methods for its search efficiency and simplicity.
We propose an alternative perturbation-based architecture selection that directly measures each operation's influence on the supernet.
We find that several failure modes of DARTS can be greatly alleviated with the proposed selection method.
arXiv Detail & Related papers (2021-08-10T00:53:39Z) - iDARTS: Differentiable Architecture Search with Stochastic Implicit
Gradients [75.41173109807735]
Differentiable ARchiTecture Search (DARTS) has recently become the mainstream of neural architecture search (NAS)
We tackle the hypergradient computation in DARTS based on the implicit function theorem.
We show that the architecture optimisation with the proposed method, named iDARTS, is expected to converge to a stationary point.
arXiv Detail & Related papers (2021-06-21T00:44:11Z) - DrNAS: Dirichlet Neural Architecture Search [88.56953713817545]
We treat the continuously relaxed architecture mixing weight as random variables, modeled by Dirichlet distribution.
With recently developed pathwise derivatives, the Dirichlet parameters can be easily optimized with gradient-based generalization.
To alleviate the large memory consumption of differentiable NAS, we propose a simple yet effective progressive learning scheme.
arXiv Detail & Related papers (2020-06-18T08:23:02Z) - Bayesian Neural Architecture Search using A Training-Free Performance
Metric [7.775212462771685]
Recurrent neural networks (RNNs) are a powerful approach for time series prediction.
This paper proposes to tackle the architecture optimization problem with a variant of the Bayesian Optimization (BO) algorithm.
Also, we propose three fixed-length encoding schemes to cope with the variable-length architecture representation.
arXiv Detail & Related papers (2020-01-29T08:42:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.