rl_reach: Reproducible Reinforcement Learning Experiments for Robotic
Reaching Tasks
- URL: http://arxiv.org/abs/2102.04916v1
- Date: Tue, 9 Feb 2021 16:14:10 GMT
- Title: rl_reach: Reproducible Reinforcement Learning Experiments for Robotic
Reaching Tasks
- Authors: Pierre Aumjaud, David McAuliffe, Francisco Javier Rodr\'iguez Lera,
Philip Cardiff
- Abstract summary: We present rl_reach, a self-contained, open-source and easy-to-use software package.
It is designed to run reproducible reinforcement learning experiments for customisable robotic reaching tasks.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Training reinforcement learning agents at solving a given task is highly
dependent on identifying optimal sets of hyperparameters and selecting suitable
environment input / output configurations. This tedious process could be eased
with a straightforward toolbox allowing its user to quickly compare different
training parameter sets. We present rl_reach, a self-contained, open-source and
easy-to-use software package designed to run reproducible reinforcement
learning experiments for customisable robotic reaching tasks. rl_reach packs
together training environments, agents, hyperparameter optimisation tools and
policy evaluation scripts, allowing its users to quickly investigate and
identify optimal training configurations. rl_reach is publicly available at
this URL: https://github.com/PierreExeter/rl_reach.
Related papers
- SERL: A Software Suite for Sample-Efficient Robotic Reinforcement
Learning [85.21378553454672]
We develop a library containing a sample efficient off-policy deep RL method, together with methods for computing rewards and resetting the environment.
We find that our implementation can achieve very efficient learning, acquiring policies for PCB board assembly, cable routing, and object relocation.
These policies achieve perfect or near-perfect success rates, extreme robustness even under perturbations, and exhibit emergent robustness recovery and correction behaviors.
arXiv Detail & Related papers (2024-01-29T10:01:10Z) - RLIF: Interactive Imitation Learning as Reinforcement Learning [56.997263135104504]
We show how off-policy reinforcement learning can enable improved performance under assumptions that are similar but potentially even more practical than those of interactive imitation learning.
Our proposed method uses reinforcement learning with user intervention signals themselves as rewards.
This relaxes the assumption that intervening experts in interactive imitation learning should be near-optimal and enables the algorithm to learn behaviors that improve over the potential suboptimal human expert.
arXiv Detail & Related papers (2023-11-21T21:05:21Z) - Learning Visual Tracking and Reaching with Deep Reinforcement Learning
on a UR10e Robotic Arm [2.2168889407389445]
Reinforcement learning algorithms provide the potential to enable robots to learn optimal solutions to complete new tasks without reprogramming them.
Current state-of-the-art in reinforcement learning relies on fast simulations and parallelization to achieve optimal performance.
This report outlines our initial research into the application of deep reinforcement learning on an industrial UR10e robot.
arXiv Detail & Related papers (2023-08-28T15:34:43Z) - Learn Fast, Segment Well: Fast Object Segmentation Learning on the iCub
Robot [20.813028212068424]
We study different techniques that allow adapting an object segmentation model in presence of novel objects or different domains.
We propose a pipeline for fast instance segmentation learning for robotic applications where data come in stream.
We benchmark the proposed pipeline on two datasets and we deploy it on a real robot, iCub humanoid.
arXiv Detail & Related papers (2022-06-27T17:14:04Z) - Accelerating Robotic Reinforcement Learning via Parameterized Action
Primitives [92.0321404272942]
Reinforcement learning can be used to build general-purpose robotic systems.
However, training RL agents to solve robotics tasks still remains challenging.
In this work, we manually specify a library of robot action primitives (RAPS), parameterized with arguments that are learned by an RL policy.
We find that our simple change to the action interface substantially improves both the learning efficiency and task performance.
arXiv Detail & Related papers (2021-10-28T17:59:30Z) - Multitask Adaptation by Retrospective Exploration with Learned World
Models [77.34726150561087]
We propose a meta-learned addressing model called RAMa that provides training samples for the MBRL agent taken from task-agnostic storage.
The model is trained to maximize the expected agent's performance by selecting promising trajectories solving prior tasks from the storage.
arXiv Detail & Related papers (2021-10-25T20:02:57Z) - Discovery of Options via Meta-Learned Subgoals [59.2160583043938]
Temporal abstractions in the form of options have been shown to help reinforcement learning (RL) agents learn faster.
We introduce a novel meta-gradient approach for discovering useful options in multi-task RL environments.
arXiv Detail & Related papers (2021-02-12T19:50:40Z) - Hyperparameter Auto-tuning in Self-Supervised Robotic Learning [12.193817049957733]
Insufficient learning (due to convergence to local optima) results in under-performing policies whilst redundant learning wastes time and resources.
We propose an auto-tuning technique based on the Evidence Lower Bound (ELBO) for self-supervised reinforcement learning.
Our method can auto-tune online and yields the best performance at a fraction of the time and computational resources.
arXiv Detail & Related papers (2020-10-16T08:58:24Z) - Solving Challenging Dexterous Manipulation Tasks With Trajectory
Optimisation and Reinforcement Learning [14.315501760755609]
Training agents to autonomously learn how to use anthropomorphic robotic hands has the potential to lead to systems capable of performing a multitude of complex manipulation tasks.
We first introduce a suite of challenging simulated manipulation tasks that current reinforcement learning and trajectory optimisation techniques find difficult.
We then introduce a simple trajectory optimisation that performs significantly better than existing methods on these environments.
arXiv Detail & Related papers (2020-09-09T13:49:52Z) - Subset Sampling For Progressive Neural Network Learning [106.12874293597754]
Progressive Neural Network Learning is a class of algorithms that incrementally construct the network's topology and optimize its parameters based on the training data.
We propose to speed up this process by exploiting subsets of training data at each incremental training step.
Experimental results in object, scene and face recognition problems demonstrate that the proposed approach speeds up the optimization procedure considerably.
arXiv Detail & Related papers (2020-02-17T18:57:33Z) - Accelerating Reinforcement Learning for Reaching using Continuous
Curriculum Learning [6.703429330486276]
We focus on accelerating reinforcement learning (RL) training and improving the performance of multi-goal reaching tasks.
Specifically, we propose a precision-based continuous curriculum learning (PCCL) method in which the requirements are gradually adjusted during the training process.
This approach is tested using a Universal Robot 5e in both simulation and real-world multi-goal reach experiments.
arXiv Detail & Related papers (2020-02-07T10:08:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.