Learning in Sparse Rewards settings through Quality-Diversity algorithms
- URL: http://arxiv.org/abs/2203.01027v1
- Date: Wed, 2 Mar 2022 11:02:34 GMT
- Title: Learning in Sparse Rewards settings through Quality-Diversity algorithms
- Authors: Giuseppe Paolo
- Abstract summary: This thesis focuses on the problem of sparse rewards with Quality-Diversity (QD) algorithms.
The first part of the thesis focuses on learning a representation of the space in which the diversity of the policies is evaluated.
The thesis continues with the introduction of the SERENE algorithm, a method that can efficiently focus on the interesting parts of the search space.
- Score: 1.4881159885040784
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the Reinforcement Learning (RL) framework, the learning is guided through
a reward signal. This means that in situations of sparse rewards the agent has
to focus on exploration, in order to discover which action, or set of actions
leads to the reward. RL agents usually struggle with this. Exploration is the
focus of Quality-Diversity (QD) methods. In this thesis, we approach the
problem of sparse rewards with these algorithms, and in particular with Novelty
Search (NS). This is a method that only focuses on the diversity of the
possible policies behaviors. The first part of the thesis focuses on learning a
representation of the space in which the diversity of the policies is
evaluated. In this regard, we propose the TAXONS algorithm, a method that
learns a low-dimensional representation of the search space through an
AutoEncoder. While effective, TAXONS still requires information on when to
capture the observation used to learn said space. For this, we study multiple
ways, and in particular the signature transform, to encode information about
the whole trajectory of observations. The thesis continues with the
introduction of the SERENE algorithm, a method that can efficiently focus on
the interesting parts of the search space. This method separates the
exploration of the search space from the exploitation of the reward through a
two-alternating-steps approach. The exploration is performed through NS. Any
discovered reward is then locally exploited through emitters. The third and
final contribution combines TAXONS and SERENE into a single approach: STAX.
Throughout this thesis, we introduce methods that lower the amount of prior
information needed in sparse rewards settings. These contributions are a
promising step towards the development of methods that can autonomously explore
and find high-performance policies in a variety of sparse rewards settings.
Related papers
- Exploration in Deep Reinforcement Learning: A Survey [4.066140143829243]
Exploration techniques are of primary importance when solving sparse reward problems.
In sparse reward problems, the reward is rare, which means that the agent will not find the reward often by acting randomly.
This review provides a comprehensive overview of existing exploration approaches.
arXiv Detail & Related papers (2022-05-02T12:03:44Z) - Discovering and Exploiting Sparse Rewards in a Learned Behavior Space [0.46736439782713946]
Learning optimal policies in sparse rewards settings is difficult as the learning agent has little to no feedback on the quality of its actions.
We introduce STAX, an algorithm designed to learn a behavior space on-the-fly and to explore it while efficiently optimizing any reward discovered.
arXiv Detail & Related papers (2021-11-02T22:21:11Z) - MURAL: Meta-Learning Uncertainty-Aware Rewards for Outcome-Driven
Reinforcement Learning [65.52675802289775]
We show that an uncertainty aware classifier can solve challenging reinforcement learning problems.
We propose a novel method for computing the normalized maximum likelihood (NML) distribution.
We show that the resulting algorithm has a number of intriguing connections to both count-based exploration methods and prior algorithms for learning reward functions.
arXiv Detail & Related papers (2021-07-15T08:19:57Z) - MADE: Exploration via Maximizing Deviation from Explored Regions [48.49228309729319]
In online reinforcement learning (RL), efficient exploration remains challenging in high-dimensional environments with sparse rewards.
We propose a new exploration approach via textitmaximizing the deviation of the occupancy of the next policy from the explored regions.
Our approach significantly improves sample efficiency over state-of-the-art methods.
arXiv Detail & Related papers (2021-06-18T17:57:00Z) - Exploration and Incentives in Reinforcement Learning [107.42240386544633]
We consider complex exploration problems, where each agent faces the same (but unknown) MDP.
Agents control the choice of policies, whereas an algorithm can only issue recommendations.
We design an algorithm which explores all reachable states in the MDP.
arXiv Detail & Related papers (2021-02-28T00:15:53Z) - Sparse Reward Exploration via Novelty Search and Emitters [55.41644538483948]
We introduce the SparsE Reward Exploration via Novelty and Emitters (SERENE) algorithm.
SERENE separates the search space exploration and reward exploitation into two alternating processes.
A meta-scheduler allocates a global computational budget by alternating between the two processes.
arXiv Detail & Related papers (2021-02-05T12:34:54Z) - Reannealing of Decaying Exploration Based On Heuristic Measure in Deep
Q-Network [82.20059754270302]
We propose an algorithm based on the idea of reannealing, that aims at encouraging exploration only when it is needed.
We perform an illustrative case study showing that it has potential to both accelerate training and obtain a better policy.
arXiv Detail & Related papers (2020-09-29T20:40:00Z) - Never Give Up: Learning Directed Exploration Strategies [63.19616370038824]
We propose a reinforcement learning agent to solve hard exploration games by learning a range of directed exploratory policies.
We construct an episodic memory-based intrinsic reward using k-nearest neighbors over the agent's recent experience to train the directed exploratory policies.
A self-supervised inverse dynamics model is used to train the embeddings of the nearest neighbour lookup, biasing the novelty signal towards what the agent can control.
arXiv Detail & Related papers (2020-02-14T13:57:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.