Mobile Robot Path Planning in Dynamic Environments through Globally
Guided Reinforcement Learning
- URL: http://arxiv.org/abs/2005.05420v2
- Date: Fri, 11 Sep 2020 21:14:15 GMT
- Title: Mobile Robot Path Planning in Dynamic Environments through Globally
Guided Reinforcement Learning
- Authors: Binyu Wang and Zhe Liu and Qingbiao Li and Amanda Prorok
- Abstract summary: We introduce a globally guided learning reinforcement approach (G2RL) to solve the multi-robot planning problem.
G2RL incorporates a novel path reward structure that generalizes to arbitrary environments.
We evaluate our method across different map types, obstacle densities and the number of robots.
- Score: 12.813442161633116
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Path planning for mobile robots in large dynamic environments is a
challenging problem, as the robots are required to efficiently reach their
given goals while simultaneously avoiding potential conflicts with other robots
or dynamic objects. In the presence of dynamic obstacles, traditional solutions
usually employ re-planning strategies, which re-call a planning algorithm to
search for an alternative path whenever the robot encounters a conflict.
However, such re-planning strategies often cause unnecessary detours. To
address this issue, we propose a learning-based technique that exploits
environmental spatio-temporal information. Different from existing
learning-based methods, we introduce a globally guided reinforcement learning
approach (G2RL), which incorporates a novel reward structure that generalizes
to arbitrary environments. We apply G2RL to solve the multi-robot path planning
problem in a fully distributed reactive manner. We evaluate our method across
different map types, obstacle densities, and the number of robots. Experimental
results show that G2RL generalizes well, outperforming existing distributed
methods, and performing very similarly to fully centralized state-of-the-art
benchmarks.
Related papers
- Commonsense Reasoning for Legged Robot Adaptation with Vision-Language Models [81.55156507635286]
Legged robots are physically capable of navigating a diverse variety of environments and overcoming a wide range of obstructions.
Current learning methods often struggle with generalization to the long tail of unexpected situations without heavy human supervision.
We propose a system, VLM-Predictive Control (VLM-PC), combining two key components that we find to be crucial for eliciting on-the-fly, adaptive behavior selection.
arXiv Detail & Related papers (2024-07-02T21:00:30Z) - Diffusion-Reinforcement Learning Hierarchical Motion Planning in Adversarial Multi-agent Games [6.532258098619471]
We focus on a motion planning task for an evasive target in a partially observable multi-agent adversarial pursuit-evasion games (PEG)
These pursuit-evasion problems are relevant to various applications, such as search and rescue operations and surveillance robots.
We propose a hierarchical architecture that integrates a high-level diffusion model to plan global paths responsive to environment data.
arXiv Detail & Related papers (2024-03-16T03:53:55Z) - Multi-Robot Path Planning Combining Heuristics and Multi-Agent
Reinforcement Learning [0.0]
In the movement process, robots need to avoid collisions with other moving robots while minimizing their travel distance.
Previous methods for this problem either continuously replan paths using search methods to avoid conflicts or choose appropriate collision avoidance strategies based on learning approaches.
We propose a path planning method, MAPPOHR, which combines a search, empirical rules, and multi-agent reinforcement learning.
arXiv Detail & Related papers (2023-06-02T05:07:37Z) - Scalable Multi-robot Motion Planning for Congested Environments With
Topological Guidance [2.846144602096543]
Multi-robot motion planning (MRMP) is the problem of finding collision-free paths for a set of robots in a continuous state space.
We extend an existing single-robot motion planning method to leverage the improved efficiency provided by topological guidance.
We demonstrate our method's ability to efficiently plan paths in complex environments with many narrow passages, scaling to robot teams of size up to 25 times larger than existing methods.
arXiv Detail & Related papers (2022-10-13T16:26:01Z) - Decentralized Global Connectivity Maintenance for Multi-Robot
Navigation: A Reinforcement Learning Approach [12.649986200029717]
This work investigates how to navigate a multi-robot team in unknown environments while maintaining connectivity.
We propose a reinforcement learning approach to develop a decentralized policy, which is shared among multiple robots.
We validate the effectiveness of the proposed approach by comparing different combinations of connectivity constraints and behavior cloning.
arXiv Detail & Related papers (2021-09-17T13:20:19Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Simultaneous Navigation and Construction Benchmarking Environments [73.0706832393065]
We need intelligent robots for mobile construction, the process of navigating in an environment and modifying its structure according to a geometric design.
In this task, a major robot vision and learning challenge is how to exactly achieve the design without GPS.
We benchmark the performance of a handcrafted policy with basic localization and planning, and state-of-the-art deep reinforcement learning methods.
arXiv Detail & Related papers (2021-03-31T00:05:54Z) - Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic
Platforms [60.59764170868101]
Reinforcement learning methods can achieve significant performance but require a large amount of training data collected on the same robotic platform.
We formulate it as a few-shot meta-learning problem where the goal is to find a model that captures the common structure shared across different robotic platforms.
We experimentally evaluate our framework on a simulated reaching and a real-robot picking task using 400 simulated robots.
arXiv Detail & Related papers (2021-03-05T14:16:20Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z) - Learning to Generalize Across Long-Horizon Tasks from Human
Demonstrations [52.696205074092006]
Generalization Through Imitation (GTI) is a two-stage offline imitation learning algorithm.
GTI exploits a structure where demonstrated trajectories for different tasks intersect at common regions of the state space.
In the first stage of GTI, we train a policy that leverages intersections to have the capacity to compose behaviors from different demonstration trajectories together.
In the second stage of GTI, we train a goal-directed agent to generalize to novel start and goal configurations.
arXiv Detail & Related papers (2020-03-13T02:25:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.