Learning to Play Table Tennis From Scratch using Muscular Robots
- URL: http://arxiv.org/abs/2006.05935v1
- Date: Wed, 10 Jun 2020 16:43:27 GMT
- Title: Learning to Play Table Tennis From Scratch using Muscular Robots
- Authors: Dieter B\"uchler, Simon Guist, Roberto Calandra, Vincent Berenz,
Bernhard Sch\"olkopf, Jan Peters
- Abstract summary: This work is the first to (a) fail-safe learn of a safety-critical dynamic task using anthropomorphic robot arms, (b) learn a precision-demanding problem with a PAM-driven system, and (c) train robots to play table tennis without real balls.
Videos and datasets are available at muscularTT.embodied.ml.
- Score: 34.34824536814943
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dynamic tasks like table tennis are relatively easy to learn for humans but
pose significant challenges to robots. Such tasks require accurate control of
fast movements and precise timing in the presence of imprecise state estimation
of the flying ball and the robot. Reinforcement Learning (RL) has shown promise
in learning of complex control tasks from data. However, applying step-based RL
to dynamic tasks on real systems is safety-critical as RL requires exploring
and failing safely for millions of time steps in high-speed regimes. In this
paper, we demonstrate that safe learning of table tennis using model-free
Reinforcement Learning can be achieved by using robot arms driven by pneumatic
artificial muscles (PAMs). Softness and back-drivability properties of PAMs
prevent the system from leaving the safe region of its state space. In this
manner, RL empowers the robot to return and smash real balls with 5 m\s and
12m\s on average to a desired landing point. Our setup allows the agent to
learn this safety-critical task (i) without safety constraints in the
algorithm, (ii) while maximizing the speed of returned balls directly in the
reward function (iii) using a stochastic policy that acts directly on the
low-level controls of the real system and (iv) trains for thousands of trials
(v) from scratch without any prior knowledge. Additionally, we present HYSR, a
practical hybrid sim and real training that avoids playing real balls during
training by randomly replaying recorded ball trajectories in simulation and
applying actions to the real robot. This work is the first to (a) fail-safe
learn of a safety-critical dynamic task using anthropomorphic robot arms, (b)
learn a precision-demanding problem with a PAM-driven system despite the
control challenges and (c) train robots to play table tennis without real
balls. Videos and datasets are available at muscularTT.embodied.ml.
Related papers
- Learning Diverse Robot Striking Motions with Diffusion Models and Kinematically Constrained Gradient Guidance [0.3613661942047476]
We develop a novel diffusion modeling approach that is offline, constraint-guided, and expressive of diverse agile behaviors.
We demonstrate the effectiveness of our approach for time-critical robotic tasks by evaluating KCGG in two challenging domains: simulated air hockey and real table tennis.
arXiv Detail & Related papers (2024-09-23T20:26:51Z) - Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control [106.32794844077534]
This paper presents a study on using deep reinforcement learning to create dynamic locomotion controllers for bipedal robots.
We develop a general control solution that can be used for a range of dynamic bipedal skills, from periodic walking and running to aperiodic jumping and standing.
This work pushes the limits of agility for bipedal robots through extensive real-world experiments.
arXiv Detail & Related papers (2024-01-30T10:48:43Z) - Robot Fine-Tuning Made Easy: Pre-Training Rewards and Policies for
Autonomous Real-World Reinforcement Learning [58.3994826169858]
We introduce RoboFuME, a reset-free fine-tuning system for robotic reinforcement learning.
Our insights are to utilize offline reinforcement learning techniques to ensure efficient online fine-tuning of a pre-trained policy.
Our method can incorporate data from an existing robot dataset and improve on a target task within as little as 3 hours of autonomous real-world experience.
arXiv Detail & Related papers (2023-10-23T17:50:08Z) - Robotic Table Tennis: A Case Study into a High Speed Learning System [30.30242337602385]
We present a real-world robotic learning system capable of hundreds of table tennis rallies with a human.
This system puts together a highly optimized perception subsystem, a high-speed low-latency robot controller, and a simulation paradigm that can prevent damage in the real world.
arXiv Detail & Related papers (2023-09-06T18:56:20Z) - Quality-Diversity Optimisation on a Physical Robot Through
Dynamics-Aware and Reset-Free Learning [4.260312058817663]
We build upon the Reset-Free QD (RF-QD) algorithm to learn controllers directly on a physical robot.
This method uses a dynamics model, learned from interactions between the robot and the environment, to predict the robot's behaviour.
RF-QD also includes a recovery policy that returns the robot to a safe zone when it has walked outside of it, allowing continuous learning.
arXiv Detail & Related papers (2023-04-24T13:24:00Z) - Hindsight States: Blending Sim and Real Task Elements for Efficient
Reinforcement Learning [61.3506230781327]
In robotics, one approach to generate training data builds on simulations based on dynamics models derived from first principles.
Here, we leverage the imbalance in complexity of the dynamics to learn more sample-efficiently.
We validate our method on several challenging simulated tasks and demonstrate that it improves learning both alone and when combined with an existing hindsight algorithm.
arXiv Detail & Related papers (2023-03-03T21:55:04Z) - Hierarchical Reinforcement Learning for Precise Soccer Shooting Skills
using a Quadrupedal Robot [76.04391023228081]
We address the problem of enabling quadrupedal robots to perform precise shooting skills in the real world using reinforcement learning.
We propose a hierarchical framework that leverages deep reinforcement learning to train a robust motion control policy.
We deploy the proposed framework on an A1 quadrupedal robot and enable it to accurately shoot the ball to random targets in the real world.
arXiv Detail & Related papers (2022-08-01T22:34:51Z) - Accelerating Robotic Reinforcement Learning via Parameterized Action
Primitives [92.0321404272942]
Reinforcement learning can be used to build general-purpose robotic systems.
However, training RL agents to solve robotics tasks still remains challenging.
In this work, we manually specify a library of robot action primitives (RAPS), parameterized with arguments that are learned by an RL policy.
We find that our simple change to the action interface substantially improves both the learning efficiency and task performance.
arXiv Detail & Related papers (2021-10-28T17:59:30Z) - Learning of Parameters in Behavior Trees for Movement Skills [0.9562145896371784]
Behavior Trees (BTs) can provide a policy representation that supports modular and composable skills.
We present a novel algorithm that can learn the parameters of a BT policy in simulation and then generalize to the physical robot without any additional training.
arXiv Detail & Related papers (2021-09-27T13:46:39Z) - RL STaR Platform: Reinforcement Learning for Simulation based Training
of Robots [3.249853429482705]
Reinforcement learning (RL) is a promising field to enhance robotic autonomy and decision making capabilities for space robotics.
This paper introduces the RL STaR platform, and how researchers can use it through a demonstration.
arXiv Detail & Related papers (2020-09-21T03:09:53Z) - Meta-Reinforcement Learning for Robotic Industrial Insertion Tasks [70.56451186797436]
We study how to use meta-reinforcement learning to solve the bulk of the problem in simulation.
We demonstrate our approach by training an agent to successfully perform challenging real-world insertion tasks.
arXiv Detail & Related papers (2020-04-29T18:00:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.