Goal-Conditioned Reinforcement Learning: Problems and Solutions
- URL: http://arxiv.org/abs/2201.08299v1
- Date: Thu, 20 Jan 2022 17:06:42 GMT
- Title: Goal-Conditioned Reinforcement Learning: Problems and Solutions
- Authors: Minghuan Liu, Menghui Zhu, Weinan Zhang
- Abstract summary: Goal-conditioned reinforcement learning (GCRL) trains an agent to achieve different goals under particular scenarios.
In this survey, we provide a comprehensive overview of the challenges and algorithms for GCRL.
- Score: 21.51237981337685
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Goal-conditioned reinforcement learning (GCRL), related to a set of complex
RL problems, trains an agent to achieve different goals under particular
scenarios. Compared to the standard RL solutions that learn a policy solely
depending on the states or observations, GCRL additionally requires the agent
to make decisions according to different goals. In this survey, we provide a
comprehensive overview of the challenges and algorithms for GCRL. Firstly, we
answer what the basic problems are studied in this field. Then, we explain how
goals are represented and present how existing solutions are designed from
different points of view. Finally, we make the conclusion and discuss potential
future prospects that recent researches focus on.
Related papers
- A Comprehensive Survey on Inverse Constrained Reinforcement Learning: Definitions, Progress and Challenges [27.681999552782372]
Inverse Constrained Reinforcement Learning (ICRL) is the task of inferring the implicit constraints followed by expert agents from their demonstration data.
This article presents a categorical survey of the latest advances in ICRL.
It serves as a comprehensive reference for machine learning researchers and practitioners, as well as starters seeking to comprehend the definitions, advancements, and important challenges in ICRL.
arXiv Detail & Related papers (2024-09-11T18:49:03Z) - Discovering Multiple Solutions from a Single Task in Offline Reinforcement Learning [51.00472376469131]
We propose an algorithm that learns multiple solutions from a single task in offline reinforcement learning.
Our experimental results show that the proposed algorithm learns multiple qualitatively and quantitatively distinctive solutions in offline RL.
arXiv Detail & Related papers (2024-06-10T03:25:49Z) - Ensemble Reinforcement Learning: A Survey [43.17635633600716]
Reinforcement Learning (RL) has emerged as a highly effective technique for addressing various scientific and applied problems.
In response, ensemble reinforcement learning (ERL), a promising approach that combines the benefits of both RL and ensemble learning (EL), has gained widespread popularity.
ERL leverages multiple models or training algorithms to comprehensively explore the problem space and possesses strong generalization capabilities.
arXiv Detail & Related papers (2023-03-05T09:26:44Z) - A Survey of Meta-Reinforcement Learning [69.76165430793571]
We cast the development of better RL algorithms as a machine learning problem itself in a process called meta-RL.
We discuss how, at a high level, meta-RL research can be clustered based on the presence of a task distribution and the learning budget available for each individual task.
We conclude by presenting the open problems on the path to making meta-RL part of the standard toolbox for a deep RL practitioner.
arXiv Detail & Related papers (2023-01-19T12:01:41Z) - Reinforcement Learning with Stepwise Fairness Constraints [50.538878453547966]
We introduce the study of reinforcement learning with stepwise fairness constraints.
We provide learning algorithms with strong theoretical guarantees in regard to policy optimality and fairness violation.
arXiv Detail & Related papers (2022-11-08T04:06:23Z) - A Survey on Deep Reinforcement Learning-based Approaches for Adaptation
and Generalization [3.307203784120634]
Deep Reinforcement Learning (DRL) aims to create intelligent agents that can learn to solve complex problems efficiently in a real-world environment.
This paper presents a survey on the recent developments in DRL-based approaches for adaptation and generalization.
arXiv Detail & Related papers (2022-02-17T04:29:08Z) - Contextualize Me -- The Case for Context in Reinforcement Learning [49.794253971446416]
Contextual Reinforcement Learning (cRL) provides a framework to model such changes in a principled manner.
We show how cRL contributes to improving zero-shot generalization in RL through meaningful benchmarks and structured reasoning about generalization tasks.
arXiv Detail & Related papers (2022-02-09T15:01:59Z) - Explainable Reinforcement Learning for Broad-XAI: A Conceptual Framework
and Survey [0.7366405857677226]
Reinforcement Learning (RL) methods provide a potential backbone for the cognitive model required for the development of Broad-XAI.
RL represents a suite of approaches that have had increasing success in solving a range of sequential decision-making problems.
This paper aims to introduce a conceptual framework, called the Causal XRL Framework (CXF), that unifies the current XRL research and uses RL as a backbone to the development of Broad-XAI.
arXiv Detail & Related papers (2021-08-20T05:18:50Z) - Q-Mixing Network for Multi-Agent Pathfinding in Partially Observable
Grid Environments [62.997667081978825]
We consider the problem of multi-agent navigation in partially observable grid environments.
We suggest utilizing the reinforcement learning approach when the agents, first, learn the policies that map observations to actions and then follow these policies to reach their goals.
arXiv Detail & Related papers (2021-08-13T09:44:47Z) - Variational Empowerment as Representation Learning for Goal-Based
Reinforcement Learning [114.07623388322048]
We discuss how the standard goal-conditioned RL (GCRL) is encapsulated by the objective variational empowerment.
Our work lays a novel foundation from which to evaluate, analyze, and develop representation learning techniques in goal-based RL.
arXiv Detail & Related papers (2021-06-02T18:12:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.