Improving the sample-efficiency of neural architecture search with
reinforcement learning
- URL: http://arxiv.org/abs/2110.06751v1
- Date: Wed, 13 Oct 2021 14:30:09 GMT
- Title: Improving the sample-efficiency of neural architecture search with
reinforcement learning
- Authors: Attila Nagy, \'Abel Boros
- Abstract summary: In this work, we would like to contribute to the area of Automated Machine Learning (AutoML)
Our focus is on one of the most promising research directions, reinforcement learning.
The validation accuracies of the child networks serve as a reward signal for training the controller.
We propose to modify this to a more modern and complex algorithm, PPO, which has demonstrated to be faster and more stable in other environments.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Designing complex architectures has been an essential cogwheel in the
revolution deep learning has brought about in the past decade. When solving
difficult problems in a datadriven manner, a well-tried approach is to take an
architecture discovered by renowned deep learning scientists as a basis (e.g.
Inception) and try to apply it to a specific problem. This might be sufficient,
but as of now, achieving very high accuracy on a complex or yet unsolved task
requires the knowledge of highly-trained deep learning experts. In this work,
we would like to contribute to the area of Automated Machine Learning (AutoML),
specifically Neural Architecture Search (NAS), which intends to make deep
learning methods available for a wider range of society by designing neural
topologies automatically. Although several different approaches exist (e.g.
gradient-based or evolutionary algorithms), our focus is on one of the most
promising research directions, reinforcement learning. In this scenario, a
recurrent neural network (controller) is trained to create problem-specific
neural network architectures (child). The validation accuracies of the child
networks serve as a reward signal for training the controller with
reinforcement learning. The basis of our proposed work is Efficient Neural
Architecture Search (ENAS), where parameter sharing is applied among the child
networks. ENAS, like many other RL-based algorithms, emphasize the learning of
child networks as increasing their convergence result in a denser reward signal
for the controller, therefore significantly reducing training times. The
controller was originally trained with REINFORCE. In our research, we propose
to modify this to a more modern and complex algorithm, PPO, which has
demonstrated to be faster and more stable in other environments. Then, we
briefly discuss and evaluate our results.
Related papers
- An Approach for Efficient Neural Architecture Search Space Definition [0.0]
We propose a novel cell-based hierarchical search space, easy to comprehend and manipulate.
The objectives of the proposed approach are to optimize the search-time and to be general enough to handle most of state of the art CNN architectures.
arXiv Detail & Related papers (2023-10-25T08:07:29Z) - Neural Architecture Search for Dense Prediction Tasks in Computer Vision [74.9839082859151]
Deep learning has led to a rising demand for neural network architecture engineering.
neural architecture search (NAS) aims at automatically designing neural network architectures in a data-driven manner rather than manually.
NAS has become applicable to a much wider range of problems in computer vision.
arXiv Detail & Related papers (2022-02-15T08:06:50Z) - Efficient Neural Architecture Search with Performance Prediction [0.0]
We use a neural architecture search to find the best network architecture for the task at hand.
Existing NAS algorithms generally evaluate the fitness of a new architecture by fully training from scratch.
An end-to-end offline performance predictor is proposed to accelerate the evaluation of sampled architectures.
arXiv Detail & Related papers (2021-08-04T05:44:16Z) - D2RL: Deep Dense Architectures in Reinforcement Learning [47.67475810050311]
We take inspiration from successful architectural choices in computer vision and generative modelling.
We investigate the use of deeper networks and dense connections for reinforcement learning on a variety of simulated robotic learning benchmark environments.
arXiv Detail & Related papers (2020-10-19T01:27:07Z) - MS-RANAS: Multi-Scale Resource-Aware Neural Architecture Search [94.80212602202518]
We propose Multi-Scale Resource-Aware Neural Architecture Search (MS-RANAS)
We employ a one-shot architecture search approach in order to obtain a reduced search cost.
We achieve state-of-the-art results in terms of accuracy-speed trade-off.
arXiv Detail & Related papers (2020-09-29T11:56:01Z) - NAS-Navigator: Visual Steering for Explainable One-Shot Deep Neural
Network Synthesis [53.106414896248246]
We present a framework that allows analysts to effectively build the solution sub-graph space and guide the network search by injecting their domain knowledge.
Applying this technique in an iterative manner allows analysts to converge to the best performing neural network architecture for a given application.
arXiv Detail & Related papers (2020-09-28T01:48:45Z) - Multi-fidelity Neural Architecture Search with Knowledge Distillation [69.09782590880367]
We propose a bayesian multi-fidelity method for neural architecture search: MF-KD.
Knowledge distillation adds to a loss function a term forcing a network to mimic some teacher network.
We show that training for a few epochs with such a modified loss function leads to a better selection of neural architectures than training for a few epochs with a logistic loss.
arXiv Detail & Related papers (2020-06-15T12:32:38Z) - Efficient Architecture Search for Continual Learning [36.998565674813285]
Continual learning with neural networks aims to learn a sequence of tasks well.
It is often confronted with three challenges: (1) overcome the catastrophic forgetting problem, (2) adapt the current network to new tasks, and (3) control its model complexity.
We propose a novel approach named as Continual Learning with Efficient Architecture Search, or CLEAS in short.
arXiv Detail & Related papers (2020-06-07T02:59:29Z) - AutoML-Zero: Evolving Machine Learning Algorithms From Scratch [76.83052807776276]
We show that it is possible to automatically discover complete machine learning algorithms just using basic mathematical operations as building blocks.
We demonstrate this by introducing a novel framework that significantly reduces human bias through a generic search space.
We believe these preliminary successes in discovering machine learning algorithms from scratch indicate a promising new direction in the field.
arXiv Detail & Related papers (2020-03-06T19:00:04Z) - Neural Architecture Search For Fault Diagnosis [6.226564415963648]
Deep learning is suitable for processing big data, and has a strong feature extraction ability to realize end-to-end fault diagnosis systems.
Neural architecture search (NAS) is developing rapidly, and is becoming one of the next directions for deep learning.
In this paper, we proposed a NAS method for fault diagnosis using reinforcement learning.
arXiv Detail & Related papers (2020-02-19T04:03:51Z) - Neuroevolution of Neural Network Architectures Using CoDeepNEAT and
Keras [0.0]
A large portion of the work involved in a machine learning project is to define the best type of algorithm to solve a given problem.
Finding the optimal network topology and configurations for a given problem is a challenge that requires domain knowledge and testing efforts.
arXiv Detail & Related papers (2020-02-11T19:03:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.