LBGP: Learning Based Goal Planning for Autonomous Following in Front
- URL: http://arxiv.org/abs/2011.03125v1
- Date: Thu, 5 Nov 2020 22:29:30 GMT
- Title: LBGP: Learning Based Goal Planning for Autonomous Following in Front
- Authors: Payam Nikdel, Richard Vaughan, Mo Chen
- Abstract summary: This paper investigates a hybrid solution which combines deep reinforcement learning (RL) and classical trajectory planning.
An autonomous robot aims to stay ahead of a person as the person freely walks around.
Our system outperforms the state-of-the-art in following ahead and is more reliable compared to end-to-end alternatives in both the simulation and real world experiments.
- Score: 16.13120109400351
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper investigates a hybrid solution which combines deep reinforcement
learning (RL) and classical trajectory planning for the following in front
application. Here, an autonomous robot aims to stay ahead of a person as the
person freely walks around. Following in front is a challenging problem as the
user's intended trajectory is unknown and needs to be estimated, explicitly or
implicitly, by the robot. In addition, the robot needs to find a feasible way
to safely navigate ahead of human trajectory. Our deep RL module implicitly
estimates human trajectory and produces short-term navigational goals to guide
the robot. These goals are used by a trajectory planner to smoothly navigate
the robot to the short-term goals, and eventually in front of the user. We
employ curriculum learning in the deep RL module to efficiently achieve a high
return. Our system outperforms the state-of-the-art in following ahead and is
more reliable compared to end-to-end alternatives in both the simulation and
real world experiments. In contrast to a pure deep RL approach, we demonstrate
zero-shot transfer of the trained policy from simulation to the real world.
Related papers
- Stabilizing Contrastive RL: Techniques for Robotic Goal Reaching from
Offline Data [101.43350024175157]
Self-supervised learning has the potential to decrease the amount of human annotation and engineering effort required to learn control strategies.
Our work builds on prior work showing that the reinforcement learning (RL) itself can be cast as a self-supervised problem.
We demonstrate that a self-supervised RL algorithm based on contrastive learning can solve real-world, image-based robotic manipulation tasks.
arXiv Detail & Related papers (2023-06-06T01:36:56Z) - Learning Bipedal Walking for Humanoids with Current Feedback [5.429166905724048]
We present an approach for overcoming the sim2real gap issue for humanoid robots arising from inaccurate torque-tracking at the actuator level.
Our approach successfully trains a unified, end-to-end policy in simulation that can be deployed on a real HRP-5P humanoid robot to achieve bipedal locomotion.
arXiv Detail & Related papers (2023-03-07T08:16:46Z) - Self-Improving Robots: End-to-End Autonomous Visuomotor Reinforcement
Learning [54.636562516974884]
In imitation and reinforcement learning, the cost of human supervision limits the amount of data that robots can be trained on.
In this work, we propose MEDAL++, a novel design for self-improving robotic systems.
The robot autonomously practices the task by learning to both do and undo the task, simultaneously inferring the reward function from the demonstrations.
arXiv Detail & Related papers (2023-03-02T18:51:38Z) - Robot Navigation with Reinforcement Learned Path Generation and
Fine-Tuned Motion Control [5.187605914580086]
We propose a novel reinforcement learning based path generation (RL-PG) approach for mobile robot navigation without a prior exploration of an unknown environment.
We deploy our model on both simulation and physical platforms and demonstrate our model performs robot navigation effectively and safely.
arXiv Detail & Related papers (2022-10-19T15:10:52Z) - Advanced Skills by Learning Locomotion and Local Navigation End-to-End [10.872193480485596]
In this work, we propose to solve the complete problem by training an end-to-end policy with deep reinforcement learning.
We demonstrate the successful deployment of policies on a real quadrupedal robot.
arXiv Detail & Related papers (2022-09-26T16:35:00Z) - A Walk in the Park: Learning to Walk in 20 Minutes With Model-Free
Reinforcement Learning [86.06110576808824]
Deep reinforcement learning is a promising approach to learning policies in uncontrolled environments.
Recent advancements in machine learning algorithms and libraries combined with a carefully tuned robot controller lead to learning quadruped in only 20 minutes in the real world.
arXiv Detail & Related papers (2022-08-16T17:37:36Z) - Verifying Learning-Based Robotic Navigation Systems [61.01217374879221]
We show how modern verification engines can be used for effective model selection.
Specifically, we use verification to detect and rule out policies that may demonstrate suboptimal behavior.
Our work is the first to demonstrate the use of verification backends for recognizing suboptimal DRL policies in real-world robots.
arXiv Detail & Related papers (2022-05-26T17:56:43Z) - Intention Aware Robot Crowd Navigation with Attention-Based Interaction
Graph [3.8461692052415137]
We study the problem of safe and intention-aware robot navigation in dense and interactive crowds.
We propose a novel recurrent graph neural network with attention mechanisms to capture heterogeneous interactions among agents.
We demonstrate that our method enables the robot to achieve good navigation performance and non-invasiveness in challenging crowd navigation scenarios.
arXiv Detail & Related papers (2022-03-03T16:26:36Z) - REvolveR: Continuous Evolutionary Models for Robot-to-robot Policy
Transfer [57.045140028275036]
We consider the problem of transferring a policy across two different robots with significantly different parameters such as kinematics and morphology.
Existing approaches that train a new policy by matching the action or state transition distribution, including imitation learning methods, fail due to optimal action and/or state distribution being mismatched in different robots.
We propose a novel method named $REvolveR$ of using continuous evolutionary models for robotic policy transfer implemented in a physics simulator.
arXiv Detail & Related papers (2022-02-10T18:50:25Z) - SABER: Data-Driven Motion Planner for Autonomously Navigating
Heterogeneous Robots [112.2491765424719]
We present an end-to-end online motion planning framework that uses a data-driven approach to navigate a heterogeneous robot team towards a global goal.
We use model predictive control (SMPC) to calculate control inputs that satisfy robot dynamics, and consider uncertainty during obstacle avoidance with chance constraints.
recurrent neural networks are used to provide a quick estimate of future state uncertainty considered in the SMPC finite-time horizon solution.
A Deep Q-learning agent is employed to serve as a high-level path planner, providing the SMPC with target positions that move the robots towards a desired global goal.
arXiv Detail & Related papers (2021-08-03T02:56:21Z) - Autonomous Navigation in Dynamic Environments: Deep Learning-Based
Approach [0.0]
This thesis studies different deep learning-based approaches, highlighting the advantages and disadvantages of each scheme.
One of the deep learning methods based on convolutional neural network (CNN) is realized by software implementations.
We propose a low-cost approach, for indoor applications such as restaurants, museums, etc, on the base of using a monocular camera instead of a laser scanner.
arXiv Detail & Related papers (2021-02-03T23:20:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.