Operation-level Progressive Differentiable Architecture Search
- URL: http://arxiv.org/abs/2302.05632v1
- Date: Sat, 11 Feb 2023 09:18:01 GMT
- Title: Operation-level Progressive Differentiable Architecture Search
- Authors: Xunyu Zhu, Jian Li, Yong Liu, Weiping Wang
- Abstract summary: We propose operation-level progressive differentiable neural architecture search (OPP-DARTS) to avoid skip connections aggregation.
Our method's performance on CIFAR-10 is superior to the architecture found by standard DARTS.
- Score: 19.214462477848535
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Differentiable Neural Architecture Search (DARTS) is becoming more and more
popular among Neural Architecture Search (NAS) methods because of its high
search efficiency and low compute cost. However, the stability of DARTS is very
inferior, especially skip connections aggregation that leads to performance
collapse. Though existing methods leverage Hessian eigenvalues to alleviate
skip connections aggregation, they make DARTS unable to explore architectures
with better performance. In the paper, we propose operation-level progressive
differentiable neural architecture search (OPP-DARTS) to avoid skip connections
aggregation and explore better architectures simultaneously. We first divide
the search process into several stages during the search phase and increase
candidate operations into the search space progressively at the beginning of
each stage. It can effectively alleviate the unfair competition between
operations during the search phase of DARTS by offsetting the inherent unfair
advantage of the skip connection over other operations. Besides, to keep the
competition between operations relatively fair and select the operation from
the candidate operations set that makes training loss of the supernet largest.
The experiment results indicate that our method is effective and efficient. Our
method's performance on CIFAR-10 is superior to the architecture found by
standard DARTS, and the transferability of our method also surpasses standard
DARTS. We further demonstrate the robustness of our method on three simple
search spaces, i.e., S2, S3, S4, and the results show us that our method is
more robust than standard DARTS. Our code is available at
https://github.com/zxunyu/OPP-DARTS.
Related papers
- $\Lambda$-DARTS: Mitigating Performance Collapse by Harmonizing
Operation Selection among Cells [11.777101481512423]
Differentiable neural architecture search (DARTS) is a popular method for neural architecture search (NAS)
We show that DARTS suffers from a specific structural flaw due to its weight-sharing framework that limits the convergence of DARTS to saturation points of the softmax function.
We propose two new regularization terms that aim to prevent performance collapse by harmonizing operation selection via aligning gradients of layers.
arXiv Detail & Related papers (2022-10-14T17:54:01Z) - Enhancing the Robustness, Efficiency, and Diversity of Differentiable
Architecture Search [25.112048502327738]
Differentiable architecture search (DARTS) has attracted much attention due to its simplicity and significant improvement in efficiency.
Many works attempt to restrict the accumulation of skip connections by indicators or manual design.
We suggest a more subtle and direct approach that removes skip connections from the operation space.
arXiv Detail & Related papers (2022-04-10T13:25:36Z) - $\beta$-DARTS: Beta-Decay Regularization for Differentiable Architecture
Search [85.84110365657455]
We propose a simple-but-efficient regularization method, termed as Beta-Decay, to regularize the DARTS-based NAS searching process.
Experimental results on NAS-Bench-201 show that our proposed method can help to stabilize the searching process and makes the searched network more transferable across different datasets.
arXiv Detail & Related papers (2022-03-03T11:47:14Z) - ZARTS: On Zero-order Optimization for Neural Architecture Search [94.41017048659664]
Differentiable architecture search (DARTS) has been a popular one-shot paradigm for NAS due to its high efficiency.
This work turns to zero-order optimization and proposes a novel NAS scheme, called ZARTS, to search without enforcing the above approximation.
In particular, results on 12 benchmarks verify the outstanding robustness of ZARTS, where the performance of DARTS collapses due to its known instability issue.
arXiv Detail & Related papers (2021-10-10T09:35:15Z) - DARTS for Inverse Problems: a Study on Hyperparameter Sensitivity [21.263326724329698]
Differentiable architecture search (DARTS) is a widely researched tool for neural architecture search.
We report the results of any DARTS-based methods from several runs along with its underlying performance statistics.
arXiv Detail & Related papers (2021-08-12T10:28:02Z) - Towards Improving the Consistency, Efficiency, and Flexibility of
Differentiable Neural Architecture Search [84.4140192638394]
Most differentiable neural architecture search methods construct a super-net for search and derive a target-net as its sub-graph for evaluation.
In this paper, we introduce EnTranNAS that is composed of Engine-cells and Transit-cells.
Our method also spares much memory and computation cost, which speeds up the search process.
arXiv Detail & Related papers (2021-01-27T12:16:47Z) - ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse
Coding [86.40042104698792]
We formulate neural architecture search as a sparse coding problem.
In experiments, our two-stage method on CIFAR-10 requires only 0.05 GPU-day for search.
Our one-stage method produces state-of-the-art performances on both CIFAR-10 and ImageNet at the cost of only evaluation time.
arXiv Detail & Related papers (2020-10-13T04:34:24Z) - DARTS-: Robustly Stepping out of Performance Collapse Without Indicators [74.21019737169675]
Differentiable architecture search suffers from long-standing performance instability.
indicators such as Hessian eigenvalues are proposed as a signal to stop searching before the performance collapses.
In this paper, we undertake a more subtle and direct approach to resolve the collapse.
arXiv Detail & Related papers (2020-09-02T12:54:13Z) - RARTS: An Efficient First-Order Relaxed Architecture Search Method [5.491655566898372]
Differentiable architecture search (DARTS) is an effective method for data-driven neural network design based on solving a bilevel optimization problem.
We formulate a single level alternative and a relaxed architecture search (RARTS) method that utilizes the whole dataset in architecture learning via both data and network splitting.
For the task of searching topological architecture, i.e., the edges and the operations, RARTS obtains a higher accuracy and 60% reduction of computational cost than second-order DARTS on CIFAR-10.
arXiv Detail & Related papers (2020-08-10T04:55:51Z) - Theory-Inspired Path-Regularized Differential Network Architecture
Search [206.93821077400733]
We study the impact of skip connections to fast network optimization and its competitive advantage over other types of operations in differential architecture search (DARTS)
We propose a theory-inspired path-regularized DARTS that consists of two key modules: (i) a differential group-structured sparse binary gate introduced for each operation to avoid unfair competition among operations, and (ii) a path-depth-wise regularization used to incite search exploration for deep architectures that converge slower than shallow ones.
arXiv Detail & Related papers (2020-06-30T05:28:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.