MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis
- URL: http://arxiv.org/abs/2208.03963v1
- Date: Mon, 8 Aug 2022 08:15:34 GMT
- Title: MetaGraspNet: A Large-Scale Benchmark Dataset for Scene-Aware
Ambidextrous Bin Picking via Physics-based Metaverse Synthesis
- Authors: Maximilian Gilles, Yuhao Chen, Tim Robin Winter, E. Zhixuan Zeng,
Alexander Wong
- Abstract summary: We introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset constructed via physics-based metaverse synthesis.
The proposed dataset contains 217k RGBD images across 82 different article types, with full annotations for object detection, amodal perception, keypoint detection, manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum gripper.
We also provide a real dataset consisting of over 2.3k fully annotated high-quality RGBD images, divided into 5 levels of difficulties and an unseen object set to evaluate different object and layout properties.
- Score: 72.85526892440251
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autonomous bin picking poses significant challenges to vision-driven robotic
systems given the complexity of the problem, ranging from various sensor
modalities, to highly entangled object layouts, to diverse item properties and
gripper types. Existing methods often address the problem from one perspective.
Diverse items and complex bin scenes require diverse picking strategies
together with advanced reasoning. As such, to build robust and effective
machine-learning algorithms for solving this complex task requires significant
amounts of comprehensive and high quality data. Collecting such data in real
world would be too expensive and time prohibitive and therefore intractable
from a scalability perspective. To tackle this big, diverse data problem, we
take inspiration from the recent rise in the concept of metaverses, and
introduce MetaGraspNet, a large-scale photo-realistic bin picking dataset
constructed via physics-based metaverse synthesis. The proposed dataset
contains 217k RGBD images across 82 different article types, with full
annotations for object detection, amodal perception, keypoint detection,
manipulation order and ambidextrous grasp labels for a parallel-jaw and vacuum
gripper. We also provide a real dataset consisting of over 2.3k fully annotated
high-quality RGBD images, divided into 5 levels of difficulties and an unseen
object set to evaluate different object and layout properties. Finally, we
conduct extensive experiments showing that our proposed vacuum seal model and
synthetic dataset achieves state-of-the-art performance and generalizes to real
world use-cases.
Related papers
- Plain-Det: A Plain Multi-Dataset Object Detector [22.848784430833835]
Plain-Det offers flexibility to accommodate new datasets, in performance across diverse datasets, and training efficiency.
We conduct extensive experiments on 13 downstream datasets and Plain-Det demonstrates strong generalization capability.
arXiv Detail & Related papers (2024-07-14T05:18:06Z) - SM$^3$: Self-Supervised Multi-task Modeling with Multi-view 2D Images
for Articulated Objects [24.737865259695006]
We propose a self-supervised interaction perception method, referred to as SM$3$, to model articulated objects.
By constructing 3D geometries and textures from the captured 2D images, SM$3$ achieves integrated optimization of movable part and joint parameters.
Evaluations demonstrate that SM$3$ surpasses existing benchmarks across various categories and objects, while its adaptability in real-world scenarios has been thoroughly validated.
arXiv Detail & Related papers (2024-01-17T11:15:09Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - High-resolution Iterative Feedback Network for Camouflaged Object
Detection [128.893782016078]
Spotting camouflaged objects that are visually assimilated into the background is tricky for object detection algorithms.
We aim to extract the high-resolution texture details to avoid the detail degradation that causes blurred vision in edges and boundaries.
We introduce a novel HitNet to refine the low-resolution representations by high-resolution features in an iterative feedback manner.
arXiv Detail & Related papers (2022-03-22T11:20:21Z) - MetaGraspNet: A Large-Scale Benchmark Dataset for Vision-driven Robotic
Grasping via Physics-based Metaverse Synthesis [78.26022688167133]
We present a large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis.
The proposed dataset contains 100,000 images and 25 different object types.
We also propose a new layout-weighted performance metric alongside the dataset for evaluating object detection and segmentation performance.
arXiv Detail & Related papers (2021-12-29T17:23:24Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - REGRAD: A Large-Scale Relational Grasp Dataset for Safe and
Object-Specific Robotic Grasping in Clutter [52.117388513480435]
We present a new dataset named regrad to sustain the modeling of relationships among objects and grasps.
Our dataset is collected in both forms of 2D images and 3D point clouds.
Users are free to import their own object models for the generation of as many data as they want.
arXiv Detail & Related papers (2021-04-29T05:31:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.