An Open-Source Multi-Goal Reinforcement Learning Environment for Robotic
Manipulation with Pybullet
- URL: http://arxiv.org/abs/2105.05985v1
- Date: Wed, 12 May 2021 21:58:57 GMT
- Title: An Open-Source Multi-Goal Reinforcement Learning Environment for Robotic
Manipulation with Pybullet
- Authors: Xintong Yang, Ze Ji, Jing Wu, Yu-Kun Lai
- Abstract summary: This work re-implements the OpenAI Gym multi-goal robotic manipulation environment, originally based on the commercial Mujoco engine, onto the open-source Pybullet engine.
We provide users with new APIs to access a joint control mode, image observations and goals with customisable camera and a built-in on-hand camera.
We also design a set of multi-step, multi-goal, long-horizon and sparse reward robotic manipulation tasks, aiming to inspire new goal-conditioned reinforcement learning algorithms for such challenges.
- Score: 38.8947981067233
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This work re-implements the OpenAI Gym multi-goal robotic manipulation
environment, originally based on the commercial Mujoco engine, onto the
open-source Pybullet engine. By comparing the performances of the Hindsight
Experience Replay-aided Deep Deterministic Policy Gradient agent on both
environments, we demonstrate our successful re-implementation of the original
environment. Besides, we provide users with new APIs to access a joint control
mode, image observations and goals with customisable camera and a built-in
on-hand camera. We further design a set of multi-step, multi-goal, long-horizon
and sparse reward robotic manipulation tasks, aiming to inspire new
goal-conditioned reinforcement learning algorithms for such challenges. We use
a simple, human-prior-based curriculum learning method to benchmark the
multi-step manipulation tasks. Discussions about future research opportunities
regarding this kind of tasks are also provided.
Related papers
- Dreamitate: Real-World Visuomotor Policy Learning via Video Generation [49.03287909942888]
We propose a visuomotor policy learning framework that fine-tunes a video diffusion model on human demonstrations of a given task.
We generate an example of an execution of the task conditioned on images of a novel scene, and use this synthesized execution directly to control the robot.
arXiv Detail & Related papers (2024-06-24T17:59:45Z) - MOKA: Open-World Robotic Manipulation through Mark-Based Visual Prompting [97.52388851329667]
We introduce Marking Open-world Keypoint Affordances (MOKA) to solve robotic manipulation tasks specified by free-form language instructions.
Central to our approach is a compact point-based representation of affordance, which bridges the VLM's predictions on observed images and the robot's actions in the physical world.
We evaluate and analyze MOKA's performance on various table-top manipulation tasks including tool use, deformable body manipulation, and object rearrangement.
arXiv Detail & Related papers (2024-03-05T18:08:45Z) - Learning to Design and Use Tools for Robotic Manipulation [21.18538869008642]
Recent techniques for jointly optimizing morphology and control via deep learning are effective at designing locomotion agents.
We propose learning a designer policy, rather than a single design.
We show that this framework is more sample efficient than prior methods in multi-goal or multi-variant settings.
arXiv Detail & Related papers (2023-11-01T18:00:10Z) - Robot Fine-Tuning Made Easy: Pre-Training Rewards and Policies for
Autonomous Real-World Reinforcement Learning [58.3994826169858]
We introduce RoboFuME, a reset-free fine-tuning system for robotic reinforcement learning.
Our insights are to utilize offline reinforcement learning techniques to ensure efficient online fine-tuning of a pre-trained policy.
Our method can incorporate data from an existing robot dataset and improve on a target task within as little as 3 hours of autonomous real-world experience.
arXiv Detail & Related papers (2023-10-23T17:50:08Z) - Generalizable Long-Horizon Manipulations with Large Language Models [91.740084601715]
This work introduces a framework harnessing the capabilities of Large Language Models (LLMs) to generate primitive task conditions for generalizable long-horizon manipulations.
We create a challenging robotic manipulation task suite based on Pybullet for long-horizon task evaluation.
arXiv Detail & Related papers (2023-10-03T17:59:46Z) - Foundational Policy Acquisition via Multitask Learning for Motor Skill Generation [0.9668407688201356]
We propose a multitask reinforcement learning algorithm for foundational policy acquisition to generate novel motor skills.
Inspired by human sensorimotor adaptation mechanisms, we aim to train encoder-decoder networks that can be commonly used to learn novel motor skills.
arXiv Detail & Related papers (2023-08-31T05:26:14Z) - Learning Hierarchical Interactive Multi-Object Search for Mobile
Manipulation [10.21450780640562]
We introduce a novel interactive multi-object search task in which a robot has to open doors to navigate rooms and search inside cabinets and drawers to find target objects.
These new challenges require combining manipulation and navigation skills in unexplored environments.
We present HIMOS, a hierarchical reinforcement learning approach that learns to compose exploration, navigation, and manipulation skills.
arXiv Detail & Related papers (2023-07-12T12:25:33Z) - Augmenting Reinforcement Learning with Behavior Primitives for Diverse
Manipulation Tasks [17.13584584844048]
This work introduces MAnipulation Primitive-augmented reinforcement LEarning (MAPLE), a learning framework that augments standard reinforcement learning algorithms with a pre-defined library of behavior primitives.
We develop a hierarchical policy that involves the primitives and instantiates their executions with input parameters.
We demonstrate that MAPLE outperforms baseline approaches by a significant margin on a suite of simulated manipulation tasks.
arXiv Detail & Related papers (2021-10-07T17:44:33Z) - COG: Connecting New Skills to Past Experience with Offline Reinforcement
Learning [78.13740204156858]
We show that we can reuse prior data to extend new skills simply through dynamic programming.
We demonstrate the effectiveness of our approach by chaining together several behaviors seen in prior datasets for solving a new task.
We train our policies in an end-to-end fashion, mapping high-dimensional image observations to low-level robot control commands.
arXiv Detail & Related papers (2020-10-27T17:57:29Z) - Learning compositional models of robot skills for task and motion
planning [39.36562555272779]
We learn to use sensorimotor primitives to solve complex long-horizon manipulation problems.
We use state-of-the-art methods for active learning and sampling.
We evaluate our approach both in simulation and in the real world through measuring the quality of the selected primitive actions.
arXiv Detail & Related papers (2020-06-08T20:45:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.