ES-ENAS: Combining Evolution Strategies with Neural Architecture Search
at No Extra Cost for Reinforcement Learning
- URL: http://arxiv.org/abs/2101.07415v1
- Date: Tue, 19 Jan 2021 02:19:05 GMT
- Title: ES-ENAS: Combining Evolution Strategies with Neural Architecture Search
at No Extra Cost for Reinforcement Learning
- Authors: Xingyou Song, Krzysztof Choromanski, Jack Parker-Holder, Yunhao Tang,
Daiyi Peng, Deepali Jain, Wenbo Gao, Aldo Pacchiano, Tamas Sarlos, Yuxiang
Yang
- Abstract summary: We introduce ES-ENAS, a simple neural architecture search (NAS) algorithm for the purpose of reinforcement learning (RL) policy design.
We achieve >90% network compression for multiple tasks, which may be special interest in mobile robotics with limited storage and computational resources.
- Score: 46.4401207304477
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce ES-ENAS, a simple neural architecture search (NAS) algorithm for
the purpose of reinforcement learning (RL) policy design, by combining
Evolutionary Strategies (ES) and Efficient NAS (ENAS) in a highly scalable and
intuitive way. Our main insight is noticing that ES is already a distributed
blackbox algorithm, and thus we may simply insert a model controller from ENAS
into the central aggregator in ES and obtain weight sharing properties for
free. By doing so, we bridge the gap from NAS research in supervised learning
settings to the reinforcement learning scenario through this relatively simple
marriage between two different lines of research, and are one of the first to
apply controller-based NAS techniques to RL. We demonstrate the utility of our
method by training combinatorial neural network architectures for RL problems
in continuous control, via edge pruning and weight sharing. We also incorporate
a wide variety of popular techniques from modern NAS literature, including
multiobjective optimization and varying controller methods, to showcase their
promise in the RL field and discuss possible extensions. We achieve >90%
network compression for multiple tasks, which may be special interest in mobile
robotics with limited storage and computational resources.
Related papers
- DNA Family: Boosting Weight-Sharing NAS with Block-Wise Supervisions [121.05720140641189]
We develop a family of models with the distilling neural architecture (DNA) techniques.
Our proposed DNA models can rate all architecture candidates, as opposed to previous works that can only access a sub- search space using algorithms.
Our models achieve state-of-the-art top-1 accuracy of 78.9% and 83.6% on ImageNet for a mobile convolutional network and a small vision transformer, respectively.
arXiv Detail & Related papers (2024-03-02T22:16:47Z) - SiGeo: Sub-One-Shot NAS via Information Theory and Geometry of Loss
Landscape [14.550053893504764]
We introduce a "sub-one-shot" paradigm that serves as a bridge between zero-shot and one-shot NAS.
In sub-one-shot NAS, the supernet is trained using only a small subset of the training data, a phase we refer to as "warm-up"
We present SiGeo, a proxy founded on a novel theoretical framework that connects the supernet warm-up with the efficacy of the proxy.
arXiv Detail & Related papers (2023-11-22T05:25:24Z) - Divide-and-Conquer the NAS puzzle in Resource Constrained Federated
Learning Systems [5.53904487910875]
We propose DC-NAS -- a divide-and-conquer approach that performs supernet-based Neural Architecture Search (NAS) in a federated system by systematically sampling the search space.
We show that our approach outperforms several sampling strategies including Hadamard sampling, where the samples are maximally separated.
arXiv Detail & Related papers (2023-05-11T20:57:29Z) - NASiam: Efficient Representation Learning using Neural Architecture
Search for Siamese Networks [76.8112416450677]
Siamese networks are one of the most trending methods to achieve self-supervised visual representation learning (SSL)
NASiam is a novel approach that uses for the first time differentiable NAS to improve the multilayer perceptron projector and predictor (encoder/predictor pair)
NASiam reaches competitive performance in both small-scale (i.e., CIFAR-10/CIFAR-100) and large-scale (i.e., ImageNet) image classification datasets while costing only a few GPU hours.
arXiv Detail & Related papers (2023-01-31T19:48:37Z) - Generalization Properties of NAS under Activation and Skip Connection
Search [66.8386847112332]
We study the generalization properties of Neural Architecture Search (NAS) under a unifying framework.
We derive the lower (and upper) bounds of the minimum eigenvalue of the Neural Tangent Kernel (NTK) under the (in)finite-width regime.
We show how the derived results can guide NAS to select the top-performing architectures, even in the case without training.
arXiv Detail & Related papers (2022-09-15T12:11:41Z) - Towards Self-supervised and Weight-preserving Neural Architecture Search [38.497608743382145]
We propose the self-supervised and weight-preserving neural architecture search (SSWP-NAS) as an extension of the current NAS framework.
Experiments show that the architectures searched by the proposed framework achieve state-of-the-art accuracy on CIFAR-10, CIFAR-100, and ImageNet datasets.
arXiv Detail & Related papers (2022-06-08T18:48:05Z) - Understanding and Accelerating Neural Architecture Search with
Training-Free and Theory-Grounded Metrics [117.4281417428145]
This work targets designing a principled and unified training-free framework for Neural Architecture Search (NAS)
NAS has been explosively studied to automate the discovery of top-performer neural networks, but suffers from heavy resource consumption and often incurs search bias due to truncated training or approximations.
We present a unified framework to understand and accelerate NAS, by disentangling "TEG" characteristics of searched networks.
arXiv Detail & Related papers (2021-08-26T17:52:07Z) - Pretraining Neural Architecture Search Controllers with Locality-based
Self-Supervised Learning [0.0]
We propose a pretraining scheme that can be applied to controller-based NAS.
Our method, locality-based self-supervised classification task, leverages the structural similarity of network architectures to obtain good architecture representations.
arXiv Detail & Related papers (2021-03-15T06:30:36Z) - Binarized Neural Architecture Search for Efficient Object Recognition [120.23378346337311]
Binarized neural architecture search (BNAS) produces extremely compressed models to reduce huge computational cost on embedded devices for edge computing.
An accuracy of $96.53%$ vs. $97.22%$ is achieved on the CIFAR-10 dataset, but with a significantly compressed model, and a $40%$ faster search than the state-of-the-art PC-DARTS.
arXiv Detail & Related papers (2020-09-08T15:51:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.