How to Train Your Robot with Deep Reinforcement Learning; Lessons We've
Learned
- URL: http://arxiv.org/abs/2102.02915v1
- Date: Thu, 4 Feb 2021 22:09:28 GMT
- Title: How to Train Your Robot with Deep Reinforcement Learning; Lessons We've
Learned
- Authors: Julian Ibarz and Jie Tan and Chelsea Finn and Mrinal Kalakrishnan and
Peter Pastor and Sergey Levine
- Abstract summary: We present a number of case studies involving robotic deep RL.
We discuss commonly perceived challenges in deep RL and how they have been addressed in these works.
We also provide an overview of other outstanding challenges, many of which are unique to the real-world robotics setting.
- Score: 111.06812202454364
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep reinforcement learning (RL) has emerged as a promising approach for
autonomously acquiring complex behaviors from low level sensor observations.
Although a large portion of deep RL research has focused on applications in
video games and simulated control, which does not connect with the constraints
of learning in real environments, deep RL has also demonstrated promise in
enabling physical robots to learn complex skills in the real world. At the same
time,real world robotics provides an appealing domain for evaluating such
algorithms, as it connects directly to how humans learn; as an embodied agent
in the real world. Learning to perceive and move in the real world presents
numerous challenges, some of which are easier to address than others, and some
of which are often not considered in RL research that focuses only on simulated
domains. In this review article, we present a number of case studies involving
robotic deep RL. Building off of these case studies, we discuss commonly
perceived challenges in deep RL and how they have been addressed in these
works. We also provide an overview of other outstanding challenges, many of
which are unique to the real-world robotics setting and are not often the focus
of mainstream RL research. Our goal is to provide a resource both for
roboticists and machine learning researchers who are interested in furthering
the progress of deep RL in the real world.
Related papers
- A Retrospective on the Robot Air Hockey Challenge: Benchmarking Robust, Reliable, and Safe Learning Techniques for Real-world Robotics [53.33976793493801]
We organized the Robot Air Hockey Challenge at the NeurIPS 2023 conference.
We focus on practical challenges in robotics, such as the sim-to-real gap, low-level control issues, safety problems, real-time requirements, and the limited availability of real-world data.
Results show that solutions combining learning-based approaches with prior knowledge outperform those relying solely on data when real-world deployment is challenging.
arXiv Detail & Related papers (2024-11-08T17:20:47Z) - Deep Reinforcement Learning for Robotics: A Survey of Real-World Successes [44.619927796194915]
Reinforcement learning (RL) has shown tremendous promise across a wide range of applications.
Robotics problems pose fundamental difficulties for the application of RL, stemming from the complexity and cost of interacting with the physical world.
This survey is designed to offer insights for both RL practitioners and roboticists toward harnessing RL's power to create generally capable real-world robotic systems.
arXiv Detail & Related papers (2024-08-07T04:35:38Z) - Stabilizing Contrastive RL: Techniques for Robotic Goal Reaching from
Offline Data [101.43350024175157]
Self-supervised learning has the potential to decrease the amount of human annotation and engineering effort required to learn control strategies.
Our work builds on prior work showing that the reinforcement learning (RL) itself can be cast as a self-supervised problem.
We demonstrate that a self-supervised RL algorithm based on contrastive learning can solve real-world, image-based robotic manipulation tasks.
arXiv Detail & Related papers (2023-06-06T01:36:56Z) - Automated Reinforcement Learning (AutoRL): A Survey and Open Problems [92.73407630874841]
Automated Reinforcement Learning (AutoRL) involves not only standard applications of AutoML but also includes additional challenges unique to RL.
We provide a common taxonomy, discuss each area in detail and pose open problems which would be of interest to researchers going forward.
arXiv Detail & Related papers (2022-01-11T12:41:43Z) - Accelerating Robotic Reinforcement Learning via Parameterized Action
Primitives [92.0321404272942]
Reinforcement learning can be used to build general-purpose robotic systems.
However, training RL agents to solve robotics tasks still remains challenging.
In this work, we manually specify a library of robot action primitives (RAPS), parameterized with arguments that are learned by an RL policy.
We find that our simple change to the action interface substantially improves both the learning efficiency and task performance.
arXiv Detail & Related papers (2021-10-28T17:59:30Z) - Robust Reinforcement Learning-based Autonomous Driving Agent for
Simulation and Real World [0.0]
We present a DRL-based algorithm that is capable of performing autonomous robot control using Deep Q-Networks (DQN)
In our approach, the agent is trained in a simulated environment and it is able to navigate both in a simulated and real-world environment.
The trained agent is able to run on limited hardware resources and its performance is comparable to state-of-the-art approaches.
arXiv Detail & Related papers (2020-09-23T15:23:54Z) - Deep Reinforcement Learning with Interactive Feedback in a Human-Robot
Environment [1.2998475032187096]
We propose a deep reinforcement learning approach with interactive feedback to learn a domestic task in a human-robot scenario.
We compare three different learning methods using a simulated robotic arm for the task of organizing different objects.
The obtained results show that a learner agent, using either agent-IDeepRL or human-IDeepRL, completes the given task earlier and has fewer mistakes compared to the autonomous DeepRL approach.
arXiv Detail & Related papers (2020-07-07T11:55:27Z) - Deep reinforcement learning from human preferences [19.871618959160692]
We explore goals defined in terms of (non-expert) human preferences between pairs of trajectory segments.
We show that this approach can effectively solve complex RL tasks without access to the reward function.
This reduces the cost of human oversight far enough that it can be practically applied to state-of-the-art RL systems.
arXiv Detail & Related papers (2017-06-12T17:23:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.