BNAS:An Efficient Neural Architecture Search Approach Using Broad
Scalable Architecture
- URL: http://arxiv.org/abs/2001.06679v5
- Date: Wed, 20 Jan 2021 07:09:11 GMT
- Title: BNAS:An Efficient Neural Architecture Search Approach Using Broad
Scalable Architecture
- Authors: Zixiang Ding, Yaran Chen, Nannan Li, Dongbin Zhao, Zhiquan Sun and
C.L. Philip Chen
- Abstract summary: We propose Broad Neural Architecture Search (BNAS) where we elaborately design broad scalable architecture dubbed Broad Convolutional Neural Network (BCNN)
BNAS delivers 0.19 days which is 2.37x less expensive than ENAS who ranks the best in reinforcement learning-based NAS approaches.
- Score: 62.587982139871976
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose Broad Neural Architecture Search (BNAS) where we
elaborately design broad scalable architecture dubbed Broad Convolutional
Neural Network (BCNN) to solve the above issue. On one hand, the proposed broad
scalable architecture has fast training speed due to its shallow topology.
Moreover, we also adopt reinforcement learning and parameter sharing used in
ENAS as the optimization strategy of BNAS. Hence, the proposed approach can
achieve higher search efficiency. On the other hand, the broad scalable
architecture extracts multi-scale features and enhancement representations, and
feeds them into global average pooling layer to yield more reasonable and
comprehensive representations. Therefore, the performance of broad scalable
architecture can be promised. In particular, we also develop two variants for
BNAS who modify the topology of BCNN. In order to verify the effectiveness of
BNAS, several experiments are performed and experimental results show that 1)
BNAS delivers 0.19 days which is 2.37x less expensive than ENAS who ranks the
best in reinforcement learning-based NAS approaches, 2) compared with
small-size (0.5 millions parameters) and medium-size (1.1 millions parameters)
models, the architecture learned by BNAS obtains state-of-the-art performance
(3.58% and 3.24% test error) on CIFAR-10, 3) the learned architecture achieves
25.3% top-1 error on ImageNet just using 3.9 millions parameters.
Related papers
- When NAS Meets Trees: An Efficient Algorithm for Neural Architecture
Search [117.89827740405694]
Key challenge in neural architecture search (NAS) is designing how to explore wisely in the huge search space.
We propose a new NAS method called TNAS (NAS with trees), which improves search efficiency by exploring only a small number of architectures.
TNAS finds the global optimal architecture on CIFAR-10 with test accuracy of 94.37% in four GPU hours in NAS-Bench-201.
arXiv Detail & Related papers (2022-04-11T07:34:21Z) - BaLeNAS: Differentiable Architecture Search via the Bayesian Learning
Rule [95.56873042777316]
Differentiable Architecture Search (DARTS) has received massive attention in recent years, mainly because it significantly reduces the computational cost.
This paper formulates the neural architecture search as a distribution learning problem through relaxing the architecture weights into Gaussian distributions.
We demonstrate how the differentiable NAS benefits from Bayesian principles, enhancing exploration and improving stability.
arXiv Detail & Related papers (2021-11-25T18:13:42Z) - Stacked BNAS: Rethinking Broad Convolutional Neural Network for Neural
Architecture Search [16.6035648938434]
We propose Stacked BNAS whose search space is a developed broad scalable architecture named Stacked BCNN, with better performance than BNAS.
On the one hand, Stacked BCNN treats mini-BCNN as the basic block to preserve comprehensive representation and deliver powerful feature extraction ability.
On the other hand, we propose Knowledge Embedding Search (KES) to learn appropriate knowledge embeddings.
arXiv Detail & Related papers (2021-11-15T12:49:27Z) - TND-NAS: Towards Non-differentiable Objectives in Progressive
Differentiable NAS Framework [6.895590095853327]
Differentiable architecture search has gradually become the mainstream research topic in the field of Neural Architecture Search (NAS)
Recent differentiable NAS also aims at further improving the search performance and reducing the GPU-memory consumption.
We propose the TND-NAS, which is with the merits of the high efficiency in differentiable NAS framework and the compatibility among non-differentiable metrics in Multi-objective NAS.
arXiv Detail & Related papers (2021-11-06T14:19:36Z) - L$^{2}$NAS: Learning to Optimize Neural Architectures via
Continuous-Action Reinforcement Learning [23.25155249879658]
Differentiable architecture search (NAS) achieved remarkable results in deep neural network design.
We show that L$2$ achieves state-of-theart results on DART201 benchmark as well as NASS and Once-for-All search policies.
arXiv Detail & Related papers (2021-09-25T19:26:30Z) - Memory-Efficient Hierarchical Neural Architecture Search for Image
Restoration [68.6505473346005]
We propose a memory-efficient hierarchical NAS HiNAS (HiNAS) for image denoising and image super-resolution tasks.
With a single GTX1080Ti GPU, it takes only about 1 hour for searching for denoising network on BSD 500 and 3.5 hours for searching for the super-resolution structure on DIV2K.
arXiv Detail & Related papers (2020-12-24T12:06:17Z) - Binarized Neural Architecture Search for Efficient Object Recognition [120.23378346337311]
Binarized neural architecture search (BNAS) produces extremely compressed models to reduce huge computational cost on embedded devices for edge computing.
An accuracy of $96.53%$ vs. $97.22%$ is achieved on the CIFAR-10 dataset, but with a significantly compressed model, and a $40%$ faster search than the state-of-the-art PC-DARTS.
arXiv Detail & Related papers (2020-09-08T15:51:23Z) - DDPNAS: Efficient Neural Architecture Search via Dynamic Distribution
Pruning [135.27931587381596]
We propose an efficient and unified NAS framework termed DDPNAS via dynamic distribution pruning.
In particular, we first sample architectures from a joint categorical distribution. Then the search space is dynamically pruned and its distribution is updated every few epochs.
With the proposed efficient network generation method, we directly obtain the optimal neural architectures on given constraints.
arXiv Detail & Related papers (2019-05-28T06:35:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.