PathRL: An End-to-End Path Generation Method for Collision Avoidance via
Deep Reinforcement Learning
- URL: http://arxiv.org/abs/2310.13295v1
- Date: Fri, 20 Oct 2023 05:55:13 GMT
- Title: PathRL: An End-to-End Path Generation Method for Collision Avoidance via
Deep Reinforcement Learning
- Authors: Wenhao Yu, Jie Peng, Quecheng Qiu, Hanyu Wang, Lu Zhang and Jianmin Ji
- Abstract summary: We propose PathRL, a novel DRL method that trains the policy to generate the navigation path for the robot.
In our experiments, PathRL achieves better success rates and reduces angular variability compared to other DRL navigation methods.
- Score: 16.397594417992483
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Robot navigation using deep reinforcement learning (DRL) has shown great
potential in improving the performance of mobile robots. Nevertheless, most
existing DRL-based navigation methods primarily focus on training a policy that
directly commands the robot with low-level controls, like linear and angular
velocities, which leads to unstable speeds and unsmooth trajectories of the
robot during the long-term execution. An alternative method is to train a DRL
policy that outputs the navigation path directly. However, two roadblocks arise
for training a DRL policy that outputs paths: (1) The action space for
potential paths often involves higher dimensions comparing to low-level
commands, which increases the difficulties of training; (2) It takes multiple
time steps to track a path instead of a single time step, which requires the
path to predicate the interactions of the robot w.r.t. the dynamic environment
in multiple time steps. This, in turn, amplifies the challenges associated with
training. In response to these challenges, we propose PathRL, a novel DRL
method that trains the policy to generate the navigation path for the robot.
Specifically, we employ specific action space discretization techniques and
tailored state space representation methods to address the associated
challenges. In our experiments, PathRL achieves better success rates and
reduces angular rotation variability compared to other DRL navigation methods,
facilitating stable and smooth robot movement. We demonstrate the competitive
edge of PathRL in both real-world scenarios and multiple challenging simulation
environments.
Related papers
- Let Hybrid A* Path Planner Obey Traffic Rules: A Deep Reinforcement Learning-Based Planning Framework [0.0]
In this work, we combine low-level algorithms such as the hybrid A* path planning with deep reinforcement learning (DRL) to make high-level decisions.
The hybrid A* planner is able to generate a collision-free trajectory to be executed by a model predictive controller (MPC)
In addition, the DRL algorithm is able to keep the lane change command consistent within a chosen time-period.
arXiv Detail & Related papers (2024-07-01T12:00:10Z) - Aquatic Navigation: A Challenging Benchmark for Deep Reinforcement Learning [53.3760591018817]
We propose a new benchmarking environment for aquatic navigation using recent advances in the integration between game engines and Deep Reinforcement Learning.
Specifically, we focus on PPO, one of the most widely accepted algorithms, and we propose advanced training techniques.
Our empirical evaluation shows that a well-designed combination of these ingredients can achieve promising results.
arXiv Detail & Related papers (2024-05-30T23:20:23Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - SERL: A Software Suite for Sample-Efficient Robotic Reinforcement
Learning [85.21378553454672]
We develop a library containing a sample efficient off-policy deep RL method, together with methods for computing rewards and resetting the environment.
We find that our implementation can achieve very efficient learning, acquiring policies for PCB board assembly, cable routing, and object relocation.
These policies achieve perfect or near-perfect success rates, extreme robustness even under perturbations, and exhibit emergent robustness recovery and correction behaviors.
arXiv Detail & Related papers (2024-01-29T10:01:10Z) - Grow Your Limits: Continuous Improvement with Real-World RL for Robotic
Locomotion [66.69666636971922]
We present APRL, a policy regularization framework that modulates the robot's exploration over the course of training.
APRL enables a quadrupedal robot to efficiently learn to walk entirely in the real world within minutes.
arXiv Detail & Related papers (2023-10-26T17:51:46Z) - Robot Navigation with Reinforcement Learned Path Generation and
Fine-Tuned Motion Control [5.187605914580086]
We propose a novel reinforcement learning based path generation (RL-PG) approach for mobile robot navigation without a prior exploration of an unknown environment.
We deploy our model on both simulation and physical platforms and demonstrate our model performs robot navigation effectively and safely.
arXiv Detail & Related papers (2022-10-19T15:10:52Z) - Accelerating Robotic Reinforcement Learning via Parameterized Action
Primitives [92.0321404272942]
Reinforcement learning can be used to build general-purpose robotic systems.
However, training RL agents to solve robotics tasks still remains challenging.
In this work, we manually specify a library of robot action primitives (RAPS), parameterized with arguments that are learned by an RL policy.
We find that our simple change to the action interface substantially improves both the learning efficiency and task performance.
arXiv Detail & Related papers (2021-10-28T17:59:30Z) - Rule-Based Reinforcement Learning for Efficient Robot Navigation with
Space Reduction [8.279526727422288]
In this paper, we focus on efficient navigation with the reinforcement learning (RL) technique.
We employ a reduction rule to shrink the trajectory, which in turn effectively reduces the redundant exploration space.
Experiments conducted on real robot navigation problems in hex-grid environments demonstrate that RuRL can achieve improved navigation performance.
arXiv Detail & Related papers (2021-04-15T07:40:27Z) - CLAMGen: Closed-Loop Arm Motion Generation via Multi-view Vision-Based
RL [4.014524824655106]
We propose a vision-based reinforcement learning (RL) approach for closed-loop trajectory generation in an arm reaching problem.
Arm trajectory generation is a fundamental robotics problem which entails finding collision-free paths to move the robot's body.
arXiv Detail & Related papers (2021-03-24T15:33:03Z) - An A* Curriculum Approach to Reinforcement Learning for RGBD Indoor
Robot Navigation [6.660458629649825]
Recently released photo-realistic simulators such as Habitat allow for the training of networks that output control actions directly from perception.
Our paper tries to overcome this problem by separating the training of the perception and control neural nets and increasing the path complexity gradually.
arXiv Detail & Related papers (2021-01-05T20:35:14Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.