Staged Reinforcement Learning for Complex Tasks through Decomposed
Environments
- URL: http://arxiv.org/abs/2311.02746v1
- Date: Sun, 5 Nov 2023 19:43:23 GMT
- Title: Staged Reinforcement Learning for Complex Tasks through Decomposed
Environments
- Authors: Rafael Pina, Corentin Artaud, Xiaolan Liu and Varuna De Silva
- Abstract summary: We discuss two methods that approximate RL problems to real problems.
In the context of traffic junction simulations, we demonstrate that, if we can decompose a complex task into multiple sub-tasks, solving these tasks first can be advantageous.
From a multi-agent perspective, we introduce a training structuring mechanism that exploits the use of experience learned under the popular paradigm called Centralised Training Decentralised Execution (CTDE)
- Score: 4.883558259729863
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reinforcement Learning (RL) is an area of growing interest in the field of
artificial intelligence due to its many notable applications in diverse fields.
Particularly within the context of intelligent vehicle control, RL has made
impressive progress. However, currently it is still in simulated controlled
environments where RL can achieve its full super-human potential. Although how
to apply simulation experience in real scenarios has been studied, how to
approximate simulated problems to the real dynamic problems is still a
challenge. In this paper, we discuss two methods that approximate RL problems
to real problems. In the context of traffic junction simulations, we
demonstrate that, if we can decompose a complex task into multiple sub-tasks,
solving these tasks first can be advantageous to help minimising possible
occurrences of catastrophic events in the complex task. From a multi-agent
perspective, we introduce a training structuring mechanism that exploits the
use of experience learned under the popular paradigm called Centralised
Training Decentralised Execution (CTDE). This experience can then be leveraged
in fully decentralised settings that are conceptually closer to real settings,
where agents often do not have access to a central oracle and must be treated
as isolated independent units. The results show that the proposed approaches
improve agents performance in complex tasks related to traffic junctions,
minimising potential safety-critical problems that might happen in these
scenarios. Although still in simulation, the investigated situations are
conceptually closer to real scenarios and thus, with these results, we intend
to motivate further research in the subject.
Related papers
- Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - Learning Curricula in Open-Ended Worlds [17.138779075998084]
This thesis develops a class of methods called Unsupervised Environment Design (UED)
Given an environment design space, UED automatically generates an infinite sequence or curriculum of training environments.
The findings in this thesis show that UED autocurricula can produce RL agents exhibiting significantly improved robustness.
arXiv Detail & Related papers (2023-12-03T16:44:00Z) - End-to-end Lidar-Driven Reinforcement Learning for Autonomous Racing [0.0]
Reinforcement Learning (RL) has emerged as a transformative approach in the domains of automation and robotics.
This study develops and trains an RL agent to navigate a racing environment solely using feedforward raw lidar and velocity data.
The agent's performance is then experimentally evaluated in a real-world racing scenario.
arXiv Detail & Related papers (2023-09-01T07:03:05Z) - Autonomous Reinforcement Learning: Formalism and Benchmarking [106.25788536376007]
Real-world embodied learning, such as that performed by humans and animals, is situated in a continual, non-episodic world.
Common benchmark tasks in RL are episodic, with the environment resetting between trials to provide the agent with multiple attempts.
This discrepancy presents a major challenge when attempting to take RL algorithms developed for episodic simulated environments and run them on real-world platforms.
arXiv Detail & Related papers (2021-12-17T16:28:06Z) - Exploratory State Representation Learning [63.942632088208505]
We propose a new approach called XSRL (eXploratory State Representation Learning) to solve the problems of exploration and SRL in parallel.
On one hand, it jointly learns compact state representations and a state transition estimator which is used to remove unexploitable information from the representations.
On the other hand, it continuously trains an inverse model, and adds to the prediction error of this model a $k$-step learning progress bonus to form the objective of a discovery policy.
arXiv Detail & Related papers (2021-09-28T10:11:07Z) - CausalCity: Complex Simulations with Agency for Causal Discovery and
Reasoning [68.74447489372037]
We present a high-fidelity simulation environment that is designed for developing algorithms for causal discovery and counterfactual reasoning.
A core component of our work is to introduce textitagency, such that it is simple to define and create complex scenarios.
We perform experiments with three state-of-the-art methods to create baselines and highlight the affordances of this environment.
arXiv Detail & Related papers (2021-06-25T00:21:41Z) - Scenic4RL: Programmatic Modeling and Generation of Reinforcement
Learning Environments [89.04823188871906]
Generation of diverse realistic scenarios is challenging for real-time strategy (RTS) environments.
Most of the existing simulators rely on randomly generating the environments.
We introduce the benefits of adopting an existing formal scenario specification language, SCENIC, to assist researchers.
arXiv Detail & Related papers (2021-06-18T21:49:46Z) - Ubiquitous Distributed Deep Reinforcement Learning at the Edge:
Analyzing Byzantine Agents in Discrete Action Spaces [0.06554326244334865]
This paper discusses some of the challenges in multi-agent distributed deep reinforcement learning that can occur in the presence of byzantine or malfunctioning agents.
We show how wrong discrete actions can significantly affect the collaborative learning effort.
Experiments are carried out in a simulation environment using the Atari testbed for the discrete action spaces, and advantage actor-critic (A2C) for the distributed multi-agent training.
arXiv Detail & Related papers (2020-08-18T11:25:39Z) - From Simulation to Real World Maneuver Execution using Deep
Reinforcement Learning [69.23334811890919]
Deep Reinforcement Learning has proved to be able to solve many control tasks in different fields, but the behavior of these systems is not always as expected when deployed in real-world scenarios.
This is mainly due to the lack of domain adaptation between simulated and real-world data together with the absence of distinction between train and test datasets.
We present a system based on multiple environments in which agents are trained simultaneously, evaluating the behavior of the model in different scenarios.
arXiv Detail & Related papers (2020-05-13T14:22:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.