Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes
- URL: http://arxiv.org/abs/2110.06192v1
- Date: Tue, 12 Oct 2021 17:46:06 GMT
- Title: Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes
- Authors: Alex X. Lee, Coline Devin, Yuxiang Zhou, Thomas Lampe, Konstantinos
Bousmalis, Jost Tobias Springenberg, Arunkumar Byravan, Abbas Abdolmaleki,
Nimrod Gileadi, David Khosid, Claudio Fantacci, Jose Enrique Chen, Akhil
Raju, Rae Jeong, Michael Neunert, Antoine Laurens, Stefano Saliceti, Federico
Casarini, Martin Riedmiller, Raia Hadsell, Francesco Nori
- Abstract summary: We study the problem of robotic stacking with objects of complex geometry.
We propose a challenging and diverse set of objects that was carefully designed to require strategies beyond a simple "pick-and-place" solution.
Our method is a reinforcement learning (RL) approach combined with vision-based interactive policy distillation and simulation-to-reality transfer.
- Score: 29.49728031012592
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of robotic stacking with objects of complex geometry. We
propose a challenging and diverse set of such objects that was carefully
designed to require strategies beyond a simple "pick-and-place" solution. Our
method is a reinforcement learning (RL) approach combined with vision-based
interactive policy distillation and simulation-to-reality transfer. Our learned
policies can efficiently handle multiple object combinations in the real world
and exhibit a large variety of stacking skills. In a large experimental study,
we investigate what choices matter for learning such general vision-based
agents in simulation, and what affects optimal transfer to the real robot. We
then leverage data collected by such policies and improve upon them with
offline RL. A video and a blog post of our work are provided as supplementary
material.
Related papers
- Part-Guided 3D RL for Sim2Real Articulated Object Manipulation [27.422878372169805]
We propose a part-guided 3D RL framework, which can learn to manipulate articulated objects without demonstrations.
We combine the strengths of 2D segmentation and 3D RL to improve the efficiency of RL policy training.
A single versatile RL policy can be trained on multiple articulated object manipulation tasks simultaneously in simulation.
arXiv Detail & Related papers (2024-04-26T10:18:17Z) - CORN: Contact-based Object Representation for Nonprehensile Manipulation of General Unseen Objects [1.3299507495084417]
Nonprehensile manipulation is essential for manipulating objects that are too thin, large, or otherwise ungraspable in the wild.
We propose a novel contact-based object representation and pretraining pipeline to tackle this.
arXiv Detail & Related papers (2024-03-16T01:47:53Z) - Nonprehensile Planar Manipulation through Reinforcement Learning with
Multimodal Categorical Exploration [8.343657309038285]
Reinforcement Learning is a powerful framework for developing such robot controllers.
We propose a multimodal exploration approach through categorical distributions, which enables us to train planar pushing RL policies.
We show that the learned policies are robust to external disturbances and observation noise, and scale to tasks with multiple pushers.
arXiv Detail & Related papers (2023-08-04T16:55:00Z) - Polybot: Training One Policy Across Robots While Embracing Variability [70.74462430582163]
We propose a set of key design decisions to train a single policy for deployment on multiple robotic platforms.
Our framework first aligns the observation and action spaces of our policy across embodiments via utilizing wrist cameras.
We evaluate our method on a dataset collected over 60 hours spanning 6 tasks and 3 robots with varying joint configurations and sizes.
arXiv Detail & Related papers (2023-07-07T17:21:16Z) - Model-Based Reinforcement Learning with Multi-Task Offline Pretraining [59.82457030180094]
We present a model-based RL method that learns to transfer potentially useful dynamics and action demonstrations from offline data to a novel task.
The main idea is to use the world models not only as simulators for behavior learning but also as tools to measure the task relevance.
We demonstrate the advantages of our approach compared with the state-of-the-art methods in Meta-World and DeepMind Control Suite.
arXiv Detail & Related papers (2023-06-06T02:24:41Z) - Co-learning Planning and Control Policies Constrained by Differentiable
Logic Specifications [4.12484724941528]
This paper presents a novel reinforcement learning approach to solving high-dimensional robot navigation tasks.
We train high-quality policies with much fewer samples compared to existing reinforcement learning algorithms.
Our approach also demonstrates capabilities for high-dimensional control and avoiding suboptimal policies via policy alignment.
arXiv Detail & Related papers (2023-03-02T15:24:24Z) - Sim2real Transfer Learning for Point Cloud Segmentation: An Industrial
Application Case on Autonomous Disassembly [55.41644538483948]
We present an industrial application case that uses sim2real transfer learning for point cloud data.
We provide insights on how to generate and process synthetic point cloud data.
A novel patch-based attention network is proposed additionally to tackle this problem.
arXiv Detail & Related papers (2023-01-12T14:00:37Z) - DeXtreme: Transfer of Agile In-hand Manipulation from Simulation to
Reality [64.51295032956118]
We train a policy that can perform robust dexterous manipulation on an anthropomorphic robot hand.
Our work reaffirms the possibilities of sim-to-real transfer for dexterous manipulation in diverse kinds of hardware and simulator setups.
arXiv Detail & Related papers (2022-10-25T01:51:36Z) - Affordance Learning from Play for Sample-Efficient Policy Learning [30.701546777177555]
We use a self-supervised visual affordance model from human teleoperated play data to enable efficient policy learning and motion planning.
We combine model-based planning with model-free deep reinforcement learning to learn policies that favor the same object regions favored by people.
We find that our policies train 4x faster than the baselines and generalize better to novel objects because our visual affordance model can anticipate their affordance regions.
arXiv Detail & Related papers (2022-03-01T11:00:35Z) - Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic
Platforms [60.59764170868101]
Reinforcement learning methods can achieve significant performance but require a large amount of training data collected on the same robotic platform.
We formulate it as a few-shot meta-learning problem where the goal is to find a model that captures the common structure shared across different robotic platforms.
We experimentally evaluate our framework on a simulated reaching and a real-robot picking task using 400 simulated robots.
arXiv Detail & Related papers (2021-03-05T14:16:20Z) - Meta-Reinforcement Learning for Robotic Industrial Insertion Tasks [70.56451186797436]
We study how to use meta-reinforcement learning to solve the bulk of the problem in simulation.
We demonstrate our approach by training an agent to successfully perform challenging real-world insertion tasks.
arXiv Detail & Related papers (2020-04-29T18:00:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.