A Fully Controllable Agent in the Path Planning using Goal-Conditioned
Reinforcement Learning
- URL: http://arxiv.org/abs/2205.09967v1
- Date: Fri, 20 May 2022 05:18:03 GMT
- Title: A Fully Controllable Agent in the Path Planning using Goal-Conditioned
Reinforcement Learning
- Authors: GyeongTaek Lee
- Abstract summary: In the path planning, the routes may vary depending on the number of variables such as that it is important for the agent to reach various goals.
I propose a novel reinforcement learning framework for a fully controllable agent in the path planning.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The aim of path planning is to reach the goal from starting point by
searching for the route of an agent. In the path planning, the routes may vary
depending on the number of variables such that it is important for the agent to
reach various goals. Numerous studies, however, have dealt with a single goal
that is predefined by the user. In the present study, I propose a novel
reinforcement learning framework for a fully controllable agent in the path
planning. To do this, I propose a bi-directional memory editing to obtain
various bi-directional trajectories of the agent, in which the behavior of the
agent and sub-goals are trained on the goal-conditioned RL. As for moving the
agent in various directions, I utilize the sub-goals dedicated network,
separated from a policy network. Lastly, I present the reward shaping to
shorten the number of steps for the agent to reach the goal. In the
experimental result, the agent was able to reach the various goals that have
never been visited by the agent in the training. We confirmed that the agent
could perform difficult missions such as a round trip and the agent used the
shorter route with the reward shaping.
Related papers
- AgentGym: Evolving Large Language Model-based Agents across Diverse Environments [116.97648507802926]
Large language models (LLMs) are considered a promising foundation to build such agents.
We take the first step towards building generally-capable LLM-based agents with self-evolution ability.
We propose AgentGym, a new framework featuring a variety of environments and tasks for broad, real-time, uni-format, and concurrent agent exploration.
arXiv Detail & Related papers (2024-06-06T15:15:41Z) - Personalized Path Recourse for Reinforcement Learning Agents [4.768286204382179]
The goal is to edit a given path of actions to achieve desired goals while ensuring a high similarity to the agent's original path.
We train a personalized recourse agent to generate such personalized paths.
The proposed method is applicable to both reinforcement learning and supervised learning settings.
arXiv Detail & Related papers (2023-12-14T08:10:57Z) - NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration [57.15811390835294]
This paper describes how we can train a single unified diffusion policy to handle both goal-directed navigation and goal-agnostic exploration.
We show that this unified policy results in better overall performance when navigating to visually indicated goals in novel environments.
Our experiments, conducted on a real-world mobile robot platform, show effective navigation in unseen environments in comparison with five alternative methods.
arXiv Detail & Related papers (2023-10-11T21:07:14Z) - Learning Graph-Enhanced Commander-Executor for Multi-Agent Navigation [28.71585436726336]
Multi-agent reinforcement learning (MARL) has shown promising results for solving this issue.
Goal-conditioned hierarchical reinforcement learning (HRL) provides a promising direction to tackle this challenge.
We propose MAGE-X, a graph-based goal-conditioned hierarchical method for multi-agent navigation tasks.
arXiv Detail & Related papers (2023-02-08T14:44:21Z) - Discrete Factorial Representations as an Abstraction for Goal
Conditioned Reinforcement Learning [99.38163119531745]
We show that applying a discretizing bottleneck can improve performance in goal-conditioned RL setups.
We experimentally prove the expected return on out-of-distribution goals, while still allowing for specifying goals with expressive structure.
arXiv Detail & Related papers (2022-11-01T03:31:43Z) - Towards Using Promises for Multi-Agent Cooperation in Goal Reasoning [15.924281804465254]
We show how promises can be incorporated into the goal life cycle, a commonly used goal refinement mechanism.
We then show how promises can be used when planning for a particular goal by connecting them to timed initial literals.
arXiv Detail & Related papers (2022-06-20T15:57:51Z) - Learning user-defined sub-goals using memory editing in reinforcement
learning [0.0]
The aim of reinforcement learning (RL) is to allow the agent to achieve the final goal.
I propose a methodology to achieve the user-defined sub-goals as well as the final goal using memory editing.
I expect that this methodology can be used in the fields that need to control the agent in a variety of scenarios.
arXiv Detail & Related papers (2022-05-01T05:19:51Z) - Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning [71.52722621691365]
Building generalizable goal-conditioned agents from rich observations is a key to reinforcement learning (RL) solving real world problems.
We propose a new form of state abstraction called goal-conditioned bisimulation.
We learn this representation using a metric form of this abstraction, and show its ability to generalize to new goals in simulation manipulation tasks.
arXiv Detail & Related papers (2022-04-27T17:00:11Z) - Explaining Reinforcement Learning Policies through Counterfactual
Trajectories [147.7246109100945]
A human developer must validate that an RL agent will perform well at test-time.
Our method conveys how the agent performs under distribution shifts by showing the agent's behavior across a wider trajectory distribution.
In a user study, we demonstrate that our method enables users to score better than baseline methods on one of two agent validation tasks.
arXiv Detail & Related papers (2022-01-29T00:52:37Z) - Automatic Curriculum Learning through Value Disagreement [95.19299356298876]
Continually solving new, unsolved tasks is the key to learning diverse behaviors.
In the multi-task domain, where an agent needs to reach multiple goals, the choice of training goals can largely affect sample efficiency.
We propose setting up an automatic curriculum for goals that the agent needs to solve.
We evaluate our method across 13 multi-goal robotic tasks and 5 navigation tasks, and demonstrate performance gains over current state-of-the-art methods.
arXiv Detail & Related papers (2020-06-17T03:58:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.