Contact-GraspNet: Efficient 6-DoF Grasp Generation in Cluttered Scenes
- URL: http://arxiv.org/abs/2103.14127v1
- Date: Thu, 25 Mar 2021 20:33:29 GMT
- Title: Contact-GraspNet: Efficient 6-DoF Grasp Generation in Cluttered Scenes
- Authors: Martin Sundermeyer, Arsalan Mousavian, Rudolph Triebel, Dieter Fox
- Abstract summary: We propose an end-to-end network that efficiently generates a distribution of 6-DoF parallel-jaw grasps.
By rooting the full 6-DoF grasp pose and width in the observed point cloud, we can reduce the dimensionality of our grasp representation to 4-DoF.
In a robotic grasping study of unseen objects in structured clutter we achieve over 90% success rate, cutting the failure rate in half compared to a recent state-of-the-art method.
- Score: 50.303361537562715
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Grasping unseen objects in unconstrained, cluttered environments is an
essential skill for autonomous robotic manipulation. Despite recent progress in
full 6-DoF grasp learning, existing approaches often consist of complex
sequential pipelines that possess several potential failure points and
run-times unsuitable for closed-loop grasping. Therefore, we propose an
end-to-end network that efficiently generates a distribution of 6-DoF
parallel-jaw grasps directly from a depth recording of a scene. Our novel grasp
representation treats 3D points of the recorded point cloud as potential grasp
contacts. By rooting the full 6-DoF grasp pose and width in the observed point
cloud, we can reduce the dimensionality of our grasp representation to 4-DoF
which greatly facilitates the learning process. Our class-agnostic approach is
trained on 17 million simulated grasps and generalizes well to real world
sensor data. In a robotic grasping study of unseen objects in structured
clutter we achieve over 90% success rate, cutting the failure rate in half
compared to a recent state-of-the-art method.
Related papers
- Self-Contrastive Forward-Forward Algorithm [3.1361717406527667]
We introduce the Self-Contrastive Forward-Forward (SCFF) method, inspired by self-supervised contrastive learning.
SCFF generates positive and negative examples applicable across different datasets.
It is the first to enable FF training of recurrent neural networks, opening the door to more complex tasks.
arXiv Detail & Related papers (2024-09-17T22:58:20Z) - Language-Driven 6-DoF Grasp Detection Using Negative Prompt Guidance [13.246380364455494]
We present a new approach for language-driven 6-DoF grasp detection in cluttered point clouds.
The proposed negative prompt strategy directs the detection process toward the desired object while steering away from unwanted ones.
Our method enables an end-to-end framework where humans can command the robot to grasp desired objects in a cluttered scene using natural language.
arXiv Detail & Related papers (2024-07-18T18:24:51Z) - IPoD: Implicit Field Learning with Point Diffusion for Generalizable 3D Object Reconstruction from Single RGB-D Images [50.4538089115248]
Generalizable 3D object reconstruction from single-view RGB-D images remains a challenging task.
We propose a novel approach, IPoD, which harmonizes implicit field learning with point diffusion.
Experiments conducted on the CO3D-v2 dataset affirm the superiority of IPoD, achieving 7.8% improvement in F-score and 28.6% in Chamfer distance over existing methods.
arXiv Detail & Related papers (2024-03-30T07:17:37Z) - GraNet: A Multi-Level Graph Network for 6-DoF Grasp Pose Generation in
Cluttered Scenes [0.5755004576310334]
GraNet is a graph-based grasp pose generation framework that translates a point cloud scene into multi-level graphs.
Our pipeline can thus characterize the spatial distribution of grasps in cluttered scenes, leading to a higher rate of effective grasping.
Our method achieves state-of-the-art performance on the large-scale GraspNet-1Billion benchmark, especially in grasping unseen objects.
arXiv Detail & Related papers (2023-12-06T08:36:29Z) - ALSO: Automotive Lidar Self-supervision by Occupancy estimation [70.70557577874155]
We propose a new self-supervised method for pre-training the backbone of deep perception models operating on point clouds.
The core idea is to train the model on a pretext task which is the reconstruction of the surface on which the 3D points are sampled.
The intuition is that if the network is able to reconstruct the scene surface, given only sparse input points, then it probably also captures some fragments of semantic information.
arXiv Detail & Related papers (2022-12-12T13:10:19Z) - Towards Confidence-guided Shape Completion for Robotic Applications [6.940242990198]
Deep learning has begun taking traction as effective means of inferring a complete 3D object representation from partial visual data.
We propose an object shape completion method based on an implicit 3D representation providing a confidence value for each reconstructed point.
We experimentally validate our approach by comparing reconstructed shapes with ground truths, and by deploying our shape completion algorithm in a robotic grasping pipeline.
arXiv Detail & Related papers (2022-09-09T13:48:24Z) - Self-Supervised Arbitrary-Scale Point Clouds Upsampling via Implicit
Neural Representation [79.60988242843437]
We propose a novel approach that achieves self-supervised and magnification-flexible point clouds upsampling simultaneously.
Experimental results demonstrate that our self-supervised learning based scheme achieves competitive or even better performance than supervised learning based state-of-the-art methods.
arXiv Detail & Related papers (2022-04-18T07:18:25Z) - Point Discriminative Learning for Unsupervised Representation Learning
on 3D Point Clouds [54.31515001741987]
We propose a point discriminative learning method for unsupervised representation learning on 3D point clouds.
We achieve this by imposing a novel point discrimination loss on the middle level and global level point features.
Our method learns powerful representations and achieves new state-of-the-art performance.
arXiv Detail & Related papers (2021-08-04T15:11:48Z) - Collision-Aware Target-Driven Object Grasping in Constrained
Environments [10.934615956723672]
We propose a novel Collision-Aware Reachability Predictor (CARP) for 6-DoF grasping systems.
The CARP learns to estimate the collision-free probabilities for grasp poses and significantly improves grasping in challenging environments.
The experiments in both simulation and the real world show that our approach achieves more than 75% grasping rate on novel objects.
arXiv Detail & Related papers (2021-04-01T21:44:07Z) - Towards Reading Beyond Faces for Sparsity-Aware 4D Affect Recognition [55.15661254072032]
We present a sparsity-aware deep network for automatic 4D facial expression recognition (FER)
We first propose a novel augmentation method to combat the data limitation problem for deep learning.
We then present a sparsity-aware deep network to compute the sparse representations of convolutional features over multi-views.
arXiv Detail & Related papers (2020-02-08T13:09:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.