Deep Reinforcement Learning Based on Local GNN for Goal-conditioned
Deformable Object Rearranging
- URL: http://arxiv.org/abs/2302.10446v1
- Date: Tue, 21 Feb 2023 05:21:26 GMT
- Title: Deep Reinforcement Learning Based on Local GNN for Goal-conditioned
Deformable Object Rearranging
- Authors: Yuhong Deng, Chongkun Xia, Xueqian Wang and Lipeng Chen
- Abstract summary: Object rearranging is one of the most common deformable manipulation tasks, where the robot needs to rearrange a deformable object into a goal configuration.
Previous studies focus on designing an expert system for each specific task by model-based or data-driven approaches.
We design a local GNN (Graph Neural Network) based learning method, which utilizes two representation graphs to encode keypoints detected from images.
Our framework is effective in multiple 1-D (rope, rope ring) and 2-D (cloth) rearranging tasks in simulation and can be easily transferred to a real robot by fine-tuning a keypoint detector
- Score: 1.807492010338763
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object rearranging is one of the most common deformable manipulation tasks,
where the robot needs to rearrange a deformable object into a goal
configuration. Previous studies focus on designing an expert system for each
specific task by model-based or data-driven approaches and the application
scenarios are therefore limited. Some research has been attempting to design a
general framework to obtain more advanced manipulation capabilities for
deformable rearranging tasks, with lots of progress achieved in simulation.
However, transferring from simulation to reality is difficult due to the
limitation of the end-to-end CNN architecture. To address these challenges, we
design a local GNN (Graph Neural Network) based learning method, which utilizes
two representation graphs to encode keypoints detected from images.
Self-attention is applied for graph updating and cross-attention is applied for
generating manipulation actions. Extensive experiments have been conducted to
demonstrate that our framework is effective in multiple 1-D (rope, rope ring)
and 2-D (cloth) rearranging tasks in simulation and can be easily transferred
to a real robot by fine-tuning a keypoint detector.
Related papers
- Cross-domain and Cross-dimension Learning for Image-to-Graph
Transformers [50.576354045312115]
Direct image-to-graph transformation is a challenging task that solves object detection and relationship prediction in a single model.
We introduce a set of methods enabling cross-domain and cross-dimension transfer learning for image-to-graph transformers.
We demonstrate our method's utility in cross-domain and cross-dimension experiments, where we pretrain our models on 2D satellite images before applying them to vastly different target domains in 2D and 3D.
arXiv Detail & Related papers (2024-03-11T10:48:56Z) - Graphical Object-Centric Actor-Critic [55.2480439325792]
We propose a novel object-centric reinforcement learning algorithm combining actor-critic and model-based approaches.
We use a transformer encoder to extract object representations and graph neural networks to approximate the dynamics of an environment.
Our algorithm performs better in a visually complex 3D robotic environment and a 2D environment with compositional structure than the state-of-the-art model-free actor-critic algorithm.
arXiv Detail & Related papers (2023-10-26T06:05:12Z) - Learning visual-based deformable object rearrangement with local graph
neural networks [4.333220038316982]
We propose a novel representation strategy that can efficiently model the deformable object states with a set of keypoints and their interactions.
We also propose a light local GNN learning to jointly model the deformable rearrangement dynamics and infer the optimal manipulation actions.
Our method reaches much higher success rates on a variety of deformable rearrangement tasks (96.3% on average) than state-of-the-art method in simulation experiments.
arXiv Detail & Related papers (2023-10-16T11:42:54Z) - Transferring Foundation Models for Generalizable Robotic Manipulation [82.12754319808197]
We propose a novel paradigm that effectively leverages language-reasoning segmentation mask generated by internet-scale foundation models.
Our approach can effectively and robustly perceive object pose and enable sample-efficient generalization learning.
Demos can be found in our submitted video, and more comprehensive ones can be found in link1 or link2.
arXiv Detail & Related papers (2023-06-09T07:22:12Z) - Graph-Transporter: A Graph-based Learning Method for Goal-Conditioned
Deformable Object Rearranging Task [1.807492010338763]
We present a novel framework, Graph-Transporter, for goal-conditioned deformable object rearranging tasks.
Our framework adopts an architecture based on Fully Convolutional Network (FCN) to output pixel-wise pick-and-place actions from only visual input.
arXiv Detail & Related papers (2023-02-21T05:21:04Z) - Efficient and Robust Training of Dense Object Nets for Multi-Object
Robot Manipulation [8.321536457963655]
We propose a framework for robust and efficient training of Dense Object Nets (DON)
We focus on training with multi-object data instead of singulated objects, combined with a well-chosen augmentation scheme.
We demonstrate the robustness and accuracy of our proposed framework on a real-world robotic grasping task.
arXiv Detail & Related papers (2022-06-24T08:24:42Z) - DiffSkill: Skill Abstraction from Differentiable Physics for Deformable
Object Manipulations with Tools [96.38972082580294]
DiffSkill is a novel framework that uses a differentiable physics simulator for skill abstraction to solve deformable object manipulation tasks.
In particular, we first obtain short-horizon skills using individual tools from a gradient-based simulator.
We then learn a neural skill abstractor from the demonstration trajectories which takes RGBD images as input.
arXiv Detail & Related papers (2022-03-31T17:59:38Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Vision-driven Robotic
Grasping via Physics-based Metaverse Synthesis [78.26022688167133]
We present a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis.
The proposed dataset contains 100,000 images and 25 different object types.
We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance.
arXiv Detail & Related papers (2021-12-29T17:23:24Z) - Efficient and Interpretable Robot Manipulation with Graph Neural
Networks [7.799182201815763]
We represent manipulation tasks as operations over graphs, using graph neural networks (GNNs)
Our formulation first transforms the environment into a graph representation, then applies a trained GNN policy to predict which object to manipulate towards which goal state.
Our GNN policies are trained using very few expert demonstrations on simple tasks, and exhibits generalization over number and configurations of objects in the environment.
We present experiments which show that a single learned GNN policy can solve a variety of blockstacking tasks in both simulation and real hardware.
arXiv Detail & Related papers (2021-02-25T21:09:12Z) - A Long Horizon Planning Framework for Manipulating Rigid Pointcloud
Objects [25.428781562909606]
We present a framework for solving long-horizon planning problems involving manipulation of rigid objects.
Our method plans in the space of object subgoals and frees the planner from reasoning about robot-object interaction dynamics.
arXiv Detail & Related papers (2020-11-16T18:59:33Z) - Learning Predictive Representations for Deformable Objects Using
Contrastive Estimation [83.16948429592621]
We propose a new learning framework that jointly optimize both the visual representation model and the dynamics model.
We show substantial improvements over standard model-based learning techniques across our rope and cloth manipulation suite.
arXiv Detail & Related papers (2020-03-11T17:55:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.