SynPick: A Dataset for Dynamic Bin Picking Scene Understanding
- URL: http://arxiv.org/abs/2107.04852v1
- Date: Sat, 10 Jul 2021 14:58:43 GMT
- Title: SynPick: A Dataset for Dynamic Bin Picking Scene Understanding
- Authors: Arul Selvam Periyasamy, Max Schwarz, and Sven Behnke
- Abstract summary: We present SynPick, a synthetic dataset for dynamic scene understanding in binpicking scenarios.
In contrast to existing datasets, our dataset is both situated in a realistic industrial application domain.
The dataset is compatible with the popular BOP dataset format.
- Score: 25.706613724135046
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present SynPick, a synthetic dataset for dynamic scene understanding in
bin-picking scenarios. In contrast to existing datasets, our dataset is both
situated in a realistic industrial application domain -- inspired by the
well-known Amazon Robotics Challenge (ARC) -- and features dynamic scenes with
authentic picking actions as chosen by our picking heuristic developed for the
ARC 2017. The dataset is compatible with the popular BOP dataset format. We
describe the dataset generation process in detail, including object arrangement
generation and manipulation simulation using the NVIDIA PhysX physics engine.
To cover a large action space, we perform untargeted and targeted picking
actions, as well as random moving actions. To establish a baseline for object
perception, a state-of-the-art pose estimation approach is evaluated on the
dataset. We demonstrate the usefulness of tracking poses during manipulation
instead of single-shot estimation even with a naive filtering approach. The
generator source code and dataset are publicly available.
Related papers
- OmniPose6D: Towards Short-Term Object Pose Tracking in Dynamic Scenes from Monocular RGB [40.62577054196799]
We introduce a large-scale synthetic dataset OmniPose6D, crafted to mirror the diversity of real-world conditions.
We present a benchmarking framework for a comprehensive comparison of pose tracking algorithms.
arXiv Detail & Related papers (2024-10-09T09:01:40Z) - TK-Planes: Tiered K-Planes with High Dimensional Feature Vectors for Dynamic UAV-based Scenes [58.180556221044235]
We present a new approach to bridge the domain gap between synthetic and real-world data for unmanned aerial vehicle (UAV)-based perception.
Our formulation is designed for dynamic scenes, consisting of small moving objects or human actions.
We evaluate its performance on challenging datasets, including Okutama Action and UG2.
arXiv Detail & Related papers (2024-05-04T21:55:33Z) - VALERIE22 -- A photorealistic, richly metadata annotated dataset of
urban environments [5.439020425819001]
The VALERIE tool pipeline is a synthetic data generator developed to contribute to the understanding of domain-specific factors.
The VALERIE22 dataset was generated with the VALERIE procedural tools pipeline providing a photorealistic sensor simulation.
The dataset provides a uniquely rich set of metadata, allowing extraction of specific scene and semantic features.
arXiv Detail & Related papers (2023-08-18T15:44:45Z) - SynTable: A Synthetic Data Generation Pipeline for Unseen Object Amodal
Instance Segmentation of Cluttered Tabletop Scenes [2.8661021832561757]
We present SynTable, a Python-based dataset generator built using NVIDIA's Isaac Sim Replicator Composer.
Our dataset generation tool can render a complex 3D scene containing object meshes, materials, textures, lighting, and backgrounds.
We demonstrate the use of a sample dataset generated using SynTable by ray tracing for training a state-of-the-art model, UOAIS-Net.
arXiv Detail & Related papers (2023-07-14T13:24:42Z) - Contrastive Lift: 3D Object Instance Segmentation by Slow-Fast
Contrastive Fusion [110.84357383258818]
We propose a novel approach to lift 2D segments to 3D and fuse them by means of a neural field representation.
The core of our approach is a slow-fast clustering objective function, which is scalable and well-suited for scenes with a large number of objects.
Our approach outperforms the state-of-the-art on challenging scenes from the ScanNet, Hypersim, and Replica datasets.
arXiv Detail & Related papers (2023-06-07T17:57:45Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - Neural-Sim: Learning to Generate Training Data with NeRF [31.81496344354997]
We present the first fully differentiable synthetic data pipeline that uses Neural Radiance Fields (NeRFs) in a closed-loop with a target application's loss function.
Our approach generates data on-demand, with no human labor, to maximize accuracy for a target task.
arXiv Detail & Related papers (2022-07-22T22:48:33Z) - Learning from Temporal Spatial Cubism for Cross-Dataset Skeleton-based
Action Recognition [88.34182299496074]
Action labels are only available on a source dataset, but unavailable on a target dataset in the training stage.
We utilize a self-supervision scheme to reduce the domain shift between two skeleton-based action datasets.
By segmenting and permuting temporal segments or human body parts, we design two self-supervised learning classification tasks.
arXiv Detail & Related papers (2022-07-17T07:05:39Z) - Salient Objects in Clutter [130.63976772770368]
This paper identifies and addresses a serious design bias of existing salient object detection (SOD) datasets.
This design bias has led to a saturation in performance for state-of-the-art SOD models when evaluated on existing datasets.
We propose a new high-quality dataset and update the previous saliency benchmark.
arXiv Detail & Related papers (2021-05-07T03:49:26Z) - Meta-Sim2: Unsupervised Learning of Scene Structure for Synthetic Data
Generation [88.04759848307687]
In Meta-Sim2, we aim to learn the scene structure in addition to parameters, which is a challenging problem due to its discrete nature.
We use Reinforcement Learning to train our model, and design a feature space divergence between our synthesized and target images that is key to successful training.
We also show that this leads to downstream improvement in the performance of an object detector trained on our generated dataset as opposed to other baseline simulation methods.
arXiv Detail & Related papers (2020-08-20T17:28:45Z) - Detection and Segmentation of Custom Objects using High Distraction
Photorealistic Synthetic Data [0.5076419064097732]
We show a straightforward and useful methodology for performing instance segmentation using synthetic data.
The goal is to achieve high performance on manually-gathered and annotated real-world data of custom objects.
This white-paper provides strong evidence that photorealistic simulated data can be used in practical real world applications.
arXiv Detail & Related papers (2020-07-28T16:33:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.