Enhanced Gradient for Differentiable Architecture Search
- URL: http://arxiv.org/abs/2103.12529v1
- Date: Tue, 23 Mar 2021 13:27:24 GMT
- Title: Enhanced Gradient for Differentiable Architecture Search
- Authors: Haichao Zhang, Kuangrong Hao, Lei Gao, Xuesong Tang, and Bing Wei
- Abstract summary: We propose a neural network architecture search algorithm aiming to simultaneously improve network performance and reduce network complexity.
The proposed framework automatically builds the network architecture at two stages: block-level search and network-level search.
Experiment results demonstrate that our method outperforms all evaluated hand-crafted networks in image classification.
- Score: 17.431144144044968
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, neural architecture search (NAS) methods have been proposed
for the automatic generation of task-oriented network architecture in image
classification. However, the architectures obtained by existing NAS approaches
are optimized only for classification performance and do not adapt to devices
with limited computational resources. To address this challenge, we propose a
neural network architecture search algorithm aiming to simultaneously improve
network performance (e.g., classification accuracy) and reduce network
complexity. The proposed framework automatically builds the network
architecture at two stages: block-level search and network-level search. At the
stage of block-level search, a relaxation method based on the gradient is
proposed, using an enhanced gradient to design high-performance and
low-complexity blocks. At the stage of network-level search, we apply an
evolutionary multi-objective algorithm to complete the automatic design from
blocks to the target network. The experiment results demonstrate that our
method outperforms all evaluated hand-crafted networks in image classification,
with an error rate of on CIFAR10 and an error rate of on CIFAR100, both at
network parameter size less than one megabit. Moreover, compared with other
neural architecture search methods, our method offers a tremendous reduction in
designed network architecture parameters.
Related papers
- EM-DARTS: Hierarchical Differentiable Architecture Search for Eye Movement Recognition [54.99121380536659]
Eye movement biometrics have received increasing attention thanks to its high secure identification.
Deep learning (DL) models have been recently successfully applied for eye movement recognition.
DL architecture still is determined by human prior knowledge.
We propose EM-DARTS, a hierarchical differentiable architecture search algorithm to automatically design the DL architecture for eye movement recognition.
arXiv Detail & Related papers (2024-09-22T13:11:08Z) - TS-ENAS:Two-Stage Evolution for Cell-based Network Architecture Search [3.267963071384687]
We propose a Two-Stage Evolution for cell-based Network Architecture Search (TS-ENAS)
In our algorithm, a new cell-based search space and an effective two-stage encoding method are designed to represent cells and neural network structures.
The experimental results show that TS-ENAS can more effectively find the neural network architecture with comparative performance.
arXiv Detail & Related papers (2023-10-14T08:02:01Z) - FlowNAS: Neural Architecture Search for Optical Flow Estimation [65.44079917247369]
We propose a neural architecture search method named FlowNAS to automatically find the better encoder architecture for flow estimation task.
Experimental results show that the discovered architecture with the weights inherited from the super-network achieves 4.67% F1-all error on KITTI.
arXiv Detail & Related papers (2022-07-04T09:05:25Z) - Search Space Adaptation for Differentiable Neural Architecture Search in
Image Classification [15.641353388251465]
Differentiable neural architecture search (NAS) has a great impact by reducing the search cost to the level of training a single network.
In this paper, we propose an adaptation scheme of the search space by introducing a search scope.
The effectiveness of proposed method is demonstrated with ProxylessNAS for the image classification task.
arXiv Detail & Related papers (2022-06-05T05:27:12Z) - Pruning-as-Search: Efficient Neural Architecture Search via Channel
Pruning and Structural Reparameterization [50.50023451369742]
Pruning-as-Search (PaS) is an end-to-end channel pruning method to search out desired sub-network automatically and efficiently.
Our proposed architecture outperforms prior arts by around $1.0%$ top-1 accuracy on ImageNet-1000 classification task.
arXiv Detail & Related papers (2022-06-02T17:58:54Z) - ISTA-NAS: Efficient and Consistent Neural Architecture Search by Sparse
Coding [86.40042104698792]
We formulate neural architecture search as a sparse coding problem.
In experiments, our two-stage method on CIFAR-10 requires only 0.05 GPU-day for search.
Our one-stage method produces state-of-the-art performances on both CIFAR-10 and ImageNet at the cost of only evaluation time.
arXiv Detail & Related papers (2020-10-13T04:34:24Z) - Neural Architecture Search as Sparse Supernet [78.09905626281046]
This paper aims at enlarging the problem of Neural Architecture Search (NAS) from Single-Path and Multi-Path Search to automated Mixed-Path Search.
We model the NAS problem as a sparse supernet using a new continuous architecture representation with a mixture of sparsity constraints.
The sparse supernet enables us to automatically achieve sparsely-mixed paths upon a compact set of nodes.
arXiv Detail & Related papers (2020-07-31T14:51:52Z) - AlphaGAN: Fully Differentiable Architecture Search for Generative
Adversarial Networks [15.740179244963116]
Generative Adversarial Networks (GANs) are formulated as minimax game problems, whereby generators attempt to approach real data distributions by virtue of adversarial learning against discriminators.
In this work, we aim to boost model learning from the perspective of network architectures, by incorporating recent progress on automated architecture search into GANs.
We propose a fully differentiable search framework for generative adversarial networks, dubbed alphaGAN.
arXiv Detail & Related papers (2020-06-16T13:27:30Z) - DC-NAS: Divide-and-Conquer Neural Architecture Search [108.57785531758076]
We present a divide-and-conquer (DC) approach to effectively and efficiently search deep neural architectures.
We achieve a $75.1%$ top-1 accuracy on the ImageNet dataset, which is higher than that of state-of-the-art methods using the same search space.
arXiv Detail & Related papers (2020-05-29T09:02:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.