SoftGym: Benchmarking Deep Reinforcement Learning for Deformable Object
Manipulation
- URL: http://arxiv.org/abs/2011.07215v2
- Date: Mon, 8 Mar 2021 04:20:49 GMT
- Title: SoftGym: Benchmarking Deep Reinforcement Learning for Deformable Object
Manipulation
- Authors: Xingyu Lin, Yufei Wang, Jake Olkin, David Held
- Abstract summary: We present SoftGym, a set of open-source simulated benchmarks for manipulating deformable objects.
We evaluate a variety of algorithms on these tasks and highlight challenges for reinforcement learning algorithms.
- Score: 15.477950393687836
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Manipulating deformable objects has long been a challenge in robotics due to
its high dimensional state representation and complex dynamics. Recent success
in deep reinforcement learning provides a promising direction for learning to
manipulate deformable objects with data driven methods. However, existing
reinforcement learning benchmarks only cover tasks with direct state
observability and simple low-dimensional dynamics or with relatively simple
image-based environments, such as those with rigid objects. In this paper, we
present SoftGym, a set of open-source simulated benchmarks for manipulating
deformable objects, with a standard OpenAI Gym API and a Python interface for
creating new environments. Our benchmark will enable reproducible research in
this important area. Further, we evaluate a variety of algorithms on these
tasks and highlight challenges for reinforcement learning algorithms, including
dealing with a state representation that has a high intrinsic dimensionality
and is partially observable. The experiments and analysis indicate the
strengths and limitations of existing methods in the context of deformable
object manipulation that can help point the way forward for future methods
development. Code and videos of the learned policies can be found on our
project website.
Related papers
- GarmentLab: A Unified Simulation and Benchmark for Garment Manipulation [12.940189262612677]
GarmentLab is a content-rich benchmark and realistic simulation designed for deformable object and garment manipulation.
Our benchmark encompasses a diverse range of garment types, robotic systems and manipulators.
We evaluate state-of-the-art vision methods, reinforcement learning, and imitation learning approaches on these tasks.
arXiv Detail & Related papers (2024-11-02T10:09:08Z) - Learning Extrinsic Dexterity with Parameterized Manipulation Primitives [8.7221770019454]
We learn a sequence of actions that utilize the environment to change the object's pose.
Our approach can control the object's state through exploiting interactions between the object, the gripper, and the environment.
We evaluate our approach on picking box-shaped objects of various weight, shape, and friction properties from a constrained table-top workspace.
arXiv Detail & Related papers (2023-10-26T21:28:23Z) - Graphical Object-Centric Actor-Critic [55.2480439325792]
We propose a novel object-centric reinforcement learning algorithm combining actor-critic and model-based approaches.
We use a transformer encoder to extract object representations and graph neural networks to approximate the dynamics of an environment.
Our algorithm performs better in a visually complex 3D robotic environment and a 2D environment with compositional structure than the state-of-the-art model-free actor-critic algorithm.
arXiv Detail & Related papers (2023-10-26T06:05:12Z) - Learning Generalizable Manipulation Policies with Object-Centric 3D
Representations [65.55352131167213]
GROOT is an imitation learning method for learning robust policies with object-centric and 3D priors.
It builds policies that generalize beyond their initial training conditions for vision-based manipulation.
GROOT's performance excels in generalization over background changes, camera viewpoint shifts, and the presence of new object instances.
arXiv Detail & Related papers (2023-10-22T18:51:45Z) - Learning visual-based deformable object rearrangement with local graph
neural networks [4.333220038316982]
We propose a novel representation strategy that can efficiently model the deformable object states with a set of keypoints and their interactions.
We also propose a light local GNN learning to jointly model the deformable rearrangement dynamics and infer the optimal manipulation actions.
Our method reaches much higher success rates on a variety of deformable rearrangement tasks (96.3% on average) than state-of-the-art method in simulation experiments.
arXiv Detail & Related papers (2023-10-16T11:42:54Z) - Learning Foresightful Dense Visual Affordance for Deformable Object
Manipulation [5.181990978232578]
We study deformable object manipulation using dense visual affordance, with generalization towards diverse states.
We propose a framework for learning this representation, with novel designs such as multi-stage stable learning and efficient self-supervised data collection without experts.
arXiv Detail & Related papers (2023-03-20T12:14:13Z) - Learning to Transfer In-Hand Manipulations Using a Greedy Shape
Curriculum [79.6027464700869]
We show that natural and robust in-hand manipulation of simple objects in a dynamic simulation can be learned from a high quality motion capture example.
We propose a simple greedy curriculum search algorithm that can successfully apply to a range of objects such as a teapot, bunny, bottle, train, and elephant.
arXiv Detail & Related papers (2023-03-14T17:08:19Z) - Efficient Representations of Object Geometry for Reinforcement Learning
of Interactive Grasping Policies [29.998917158604694]
We present a reinforcement learning framework that learns the interactive grasping of various geometrically distinct real-world objects.
Videos of learned interactive policies are available at https://maltemosbach.org/io/geometry_aware_grasping_policies.
arXiv Detail & Related papers (2022-11-20T11:47:33Z) - Robust and Controllable Object-Centric Learning through Energy-based
Models [95.68748828339059]
ours is a conceptually simple and general approach to learning object-centric representations through an energy-based model.
We show that ours can be easily integrated into existing architectures and can effectively extract high-quality object-centric representations.
arXiv Detail & Related papers (2022-10-11T15:11:15Z) - ACID: Action-Conditional Implicit Visual Dynamics for Deformable Object
Manipulation [135.10594078615952]
We introduce ACID, an action-conditional visual dynamics model for volumetric deformable objects.
A benchmark contains over 17,000 action trajectories with six types of plush toys and 78 variants.
Our model achieves the best performance in geometry, correspondence, and dynamics predictions.
arXiv Detail & Related papers (2022-03-14T04:56:55Z) - Learning Predictive Representations for Deformable Objects Using
Contrastive Estimation [83.16948429592621]
We propose a new learning framework that jointly optimize both the visual representation model and the dynamics model.
We show substantial improvements over standard model-based learning techniques across our rope and cloth manipulation suite.
arXiv Detail & Related papers (2020-03-11T17:55:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.