DiffSkill: Skill Abstraction from Differentiable Physics for Deformable
Object Manipulations with Tools
- URL: http://arxiv.org/abs/2203.17275v1
- Date: Thu, 31 Mar 2022 17:59:38 GMT
- Title: DiffSkill: Skill Abstraction from Differentiable Physics for Deformable
Object Manipulations with Tools
- Authors: Xingyu Lin, Zhiao Huang, Yunzhu Li, Joshua B. Tenenbaum, David Held,
Chuang Gan
- Abstract summary: DiffSkill is a novel framework that uses a differentiable physics simulator for skill abstraction to solve deformable object manipulation tasks.
In particular, we first obtain short-horizon skills using individual tools from a gradient-based simulator.
We then learn a neural skill abstractor from the demonstration trajectories which takes RGBD images as input.
- Score: 96.38972082580294
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: We consider the problem of sequential robotic manipulation of deformable
objects using tools. Previous works have shown that differentiable physics
simulators provide gradients to the environment state and help trajectory
optimization to converge orders of magnitude faster than model-free
reinforcement learning algorithms for deformable object manipulation. However,
such gradient-based trajectory optimization typically requires access to the
full simulator states and can only solve short-horizon, single-skill tasks due
to local optima. In this work, we propose a novel framework, named DiffSkill,
that uses a differentiable physics simulator for skill abstraction to solve
long-horizon deformable object manipulation tasks from sensory observations. In
particular, we first obtain short-horizon skills using individual tools from a
gradient-based optimizer, using the full state information in a differentiable
simulator; we then learn a neural skill abstractor from the demonstration
trajectories which takes RGBD images as input. Finally, we plan over the skills
by finding the intermediate goals and then solve long-horizon tasks. We show
the advantages of our method in a new set of sequential deformable object
manipulation tasks compared to previous reinforcement learning algorithms and
compared to the trajectory optimizer.
Related papers
- Unlearning as multi-task optimization: A normalized gradient difference approach with an adaptive learning rate [105.86576388991713]
We introduce a normalized gradient difference (NGDiff) algorithm, enabling us to have better control over the trade-off between the objectives.
We provide a theoretical analysis and empirically demonstrate the superior performance of NGDiff among state-of-the-art unlearning methods on the TOFU and MUSE datasets.
arXiv Detail & Related papers (2024-10-29T14:41:44Z) - DiffVL: Scaling Up Soft Body Manipulation using Vision-Language Driven
Differentiable Physics [69.6158232150048]
DiffVL is a method that enables non-expert users to communicate soft-body manipulation tasks.
We leverage large language models to translate task descriptions into machine-interpretable optimization objectives.
arXiv Detail & Related papers (2023-12-11T14:29:25Z) - Imitation Learning via Differentiable Physics [26.356669151969953]
Imitation learning (IL) methods such as inverse reinforcement learning (IRL) usually have a double-loop training process.
We propose a new IL method, i.e., Imitation Learning via Differentiable Physics (ILD), which gets rid of the double-loop design.
ILD achieves significant improvements in final performance, convergence speed, and stability.
arXiv Detail & Related papers (2022-06-10T04:54:32Z) - Gradient-Based Trajectory Optimization With Learned Dynamics [80.41791191022139]
We use machine learning techniques to learn a differentiable dynamics model of the system from data.
We show that a neural network can model highly nonlinear behaviors accurately for large time horizons.
In our hardware experiments, we demonstrate that our learned model can represent complex dynamics for both the Spot and Radio-controlled (RC) car.
arXiv Detail & Related papers (2022-04-09T22:07:34Z) - DiffCloud: Real-to-Sim from Point Clouds with Differentiable Simulation
and Rendering of Deformable Objects [18.266002992029716]
Research in manipulation of deformable objects is typically conducted on a limited range of scenarios.
Realistic simulators with support for various types of deformations and interactions have the potential to speed up experimentation.
For highly deformable objects it is challenging to align the output of a simulator with the behavior of real objects.
arXiv Detail & Related papers (2022-04-07T00:45:26Z) - Physical Design using Differentiable Learned Simulators [9.380022457753938]
In inverse design, learned forward simulators are combined with gradient-based design optimization.
This framework produces high-quality designs by propagating through trajectories of hundreds of steps.
Our results suggest that despite some remaining challenges, machine learning-based simulators are maturing to the point where they can support general-purpose design optimization.
arXiv Detail & Related papers (2022-02-01T19:56:39Z) - Efficient Differentiable Simulation of Articulated Bodies [89.64118042429287]
We present a method for efficient differentiable simulation of articulated bodies.
This enables integration of articulated body dynamics into deep learning frameworks.
We show that reinforcement learning with articulated systems can be accelerated using gradients provided by our method.
arXiv Detail & Related papers (2021-09-16T04:48:13Z) - PlasticineLab: A Soft-Body Manipulation Benchmark with Differentiable
Physics [89.81550748680245]
We introduce a new differentiable physics benchmark called PasticineLab.
In each task, the agent uses manipulators to deform the plasticine into the desired configuration.
We evaluate several existing reinforcement learning (RL) methods and gradient-based methods on this benchmark.
arXiv Detail & Related papers (2021-04-07T17:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.