Stretchable Cells Help DARTS Search Better
- URL: http://arxiv.org/abs/2011.09300v2
- Date: Mon, 18 Apr 2022 13:43:38 GMT
- Title: Stretchable Cells Help DARTS Search Better
- Authors: Tao Huang, Shan You, Yibo Yang, Zhuozhuo Tu, Fei Wang, Chen Qian,
Changshui Zhang
- Abstract summary: Differentiable neural architecture search (DARTS) has gained much success in discovering flexible and diverse cell types.
Current DARTS methods are prone to wide and shallow cells, and this topology collapse induces sub-optimal searched cells.
In this paper, we endowing the cells with explicit stretchability, so the search can be directly implemented on our stretchable cells.
- Score: 70.52254306274092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Differentiable neural architecture search (DARTS) has gained much success in
discovering flexible and diverse cell types. To reduce the evaluation gap, the
supernet is expected to have identical layers with the target network. However,
even for this consistent search, the searched cells often suffer from poor
performance, especially for the supernet with fewer layers, as current DARTS
methods are prone to wide and shallow cells, and this topology collapse induces
sub-optimal searched cells. In this paper, we alleviate this issue by endowing
the cells with explicit stretchability, so the search can be directly
implemented on our stretchable cells for both operation type and topology
simultaneously. Concretely, we introduce a set of topological variables and a
combinatorial probabilistic distribution to explicitly model the target
topology. With more diverse and complex topologies, our method adapts well for
various layer numbers. Extensive experiments on CIFAR-10 and ImageNet show that
our stretchable cells obtain better performance with fewer layers and
parameters. For example, our method can improve DARTS by 0.28\% accuracy on
CIFAR-10 dataset with 45\% parameters reduced or 2.9\% with similar FLOPs on
ImageNet dataset.
Related papers
- TS-ENAS:Two-Stage Evolution for Cell-based Network Architecture Search [3.267963071384687]
We propose a Two-Stage Evolution for cell-based Network Architecture Search (TS-ENAS)
In our algorithm, a new cell-based search space and an effective two-stage encoding method are designed to represent cells and neural network structures.
The experimental results show that TS-ENAS can more effectively find the neural network architecture with comparative performance.
arXiv Detail & Related papers (2023-10-14T08:02:01Z) - $\Lambda$-DARTS: Mitigating Performance Collapse by Harmonizing
Operation Selection among Cells [11.777101481512423]
Differentiable neural architecture search (DARTS) is a popular method for neural architecture search (NAS)
We show that DARTS suffers from a specific structural flaw due to its weight-sharing framework that limits the convergence of DARTS to saturation points of the softmax function.
We propose two new regularization terms that aim to prevent performance collapse by harmonizing operation selection via aligning gradients of layers.
arXiv Detail & Related papers (2022-10-14T17:54:01Z) - D-DARTS: Distributed Differentiable Architecture Search [75.12821786565318]
Differentiable ARchiTecture Search (DARTS) is one of the most trending Neural Architecture Search (NAS) methods.
We propose D-DARTS, a novel solution that addresses this problem by nesting several neural networks at cell-level.
arXiv Detail & Related papers (2021-08-20T09:07:01Z) - Towards Improving the Consistency, Efficiency, and Flexibility of
Differentiable Neural Architecture Search [84.4140192638394]
Most differentiable neural architecture search methods construct a super-net for search and derive a target-net as its sub-graph for evaluation.
In this paper, we introduce EnTranNAS that is composed of Engine-cells and Transit-cells.
Our method also spares much memory and computation cost, which speeds up the search process.
arXiv Detail & Related papers (2021-01-27T12:16:47Z) - ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse
Coding [86.40042104698792]
We formulate neural architecture search as a sparse coding problem.
In experiments, our two-stage method on CIFAR-10 requires only 0.05 GPU-day for search.
Our one-stage method produces state-of-the-art performances on both CIFAR-10 and ImageNet at the cost of only evaluation time.
arXiv Detail & Related papers (2020-10-13T04:34:24Z) - DOTS: Decoupling Operation and Topology in Differentiable Architecture
Search [115.89211594258573]
Differentiable Architecture Search (DARTS) has attracted extensive attention due to its efficiency in searching for cell structures.
We propose to Decouple the Operation and Topology Search (DOTS) to make an explicit topology search.
arXiv Detail & Related papers (2020-10-02T13:00:18Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z) - Learning to segment clustered amoeboid cells from brightfield microscopy
via multi-task learning with adaptive weight selection [6.836162272841265]
We introduce a novel supervised technique for cell segmentation in a multi-task learning paradigm.
A combination of a multi-task loss, based on the region and cell boundary detection, is employed for an improved prediction efficiency of the network.
We observe an overall Dice score of 0.93 on the validation set, which is an improvement of over 15.9% on a recent unsupervised method, and outperforms the popular supervised U-net algorithm by at least $5.8%$ on average.
arXiv Detail & Related papers (2020-05-19T11:31:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.