Towards Deep Learning-based 6D Bin Pose Estimation in 3D Scans
- URL: http://arxiv.org/abs/2112.09598v1
- Date: Fri, 17 Dec 2021 16:19:06 GMT
- Title: Towards Deep Learning-based 6D Bin Pose Estimation in 3D Scans
- Authors: Luk\'a\v{s} Gajdo\v{s}ech, Viktor Kocur, Martin Stuchl\'ik,
Luk\'a\v{s} Hudec, Martin Madaras
- Abstract summary: This paper focuses on a specific task of 6D pose estimation of a bin in 3D scans.
We present a high-quality dataset composed of synthetic data and real scans captured by a structured-light scanner with precise annotations.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: An automated robotic system needs to be as robust as possible and fail-safe
in general while having relatively high precision and repeatability. Although
deep learning-based methods are becoming research standard on how to approach
3D scan and image processing tasks, the industry standard for processing this
data is still analytically-based. Our paper claims that analytical methods are
less robust and harder for testing, updating, and maintaining. This paper
focuses on a specific task of 6D pose estimation of a bin in 3D scans.
Therefore, we present a high-quality dataset composed of synthetic data and
real scans captured by a structured-light scanner with precise annotations.
Additionally, we propose two different methods for 6D bin pose estimation, an
analytical method as the industrial standard and a baseline data-driven method.
Both approaches are cross-evaluated, and our experiments show that augmenting
the training on real scans with synthetic data improves our proposed
data-driven neural model. This position paper is preliminary, as proposed
methods are trained and evaluated on a relatively small initial dataset which
we plan to extend in the future.
Related papers
- SCoDA: Domain Adaptive Shape Completion for Real Scans [78.92028595499245]
3D shape completion from point clouds is a challenging task, especially from scans of real-world objects.
We propose a new task, SCoDA, for the domain adaptation of real scan shape completion from synthetic data.
We propose a novel cross-domain feature fusion method for knowledge transfer and a novel volume-consistent self-training framework for robust learning from real data.
arXiv Detail & Related papers (2023-04-20T09:38:26Z) - What Stops Learning-based 3D Registration from Working in the Real
World? [53.68326201131434]
This work identifies the sources of 3D point cloud registration failures, analyze the reasons behind them, and propose solutions.
Ultimately, this translates to a best-practice 3D registration network (BPNet), constituting the first learning-based method able to handle previously-unseen objects in real-world data.
Our model generalizes to real data without any fine-tuning, reaching an accuracy of up to 67% on point clouds of unseen objects obtained with a commercial sensor.
arXiv Detail & Related papers (2021-11-19T19:24:27Z) - RandomRooms: Unsupervised Pre-training from Synthetic Shapes and
Randomized Layouts for 3D Object Detection [138.2892824662943]
A promising solution is to make better use of the synthetic dataset, which consists of CAD object models, to boost the learning on real datasets.
Recent work on 3D pre-training exhibits failure when transfer features learned on synthetic objects to other real-world applications.
In this work, we put forward a new method called RandomRooms to accomplish this objective.
arXiv Detail & Related papers (2021-08-17T17:56:12Z) - Secrets of 3D Implicit Object Shape Reconstruction in the Wild [92.5554695397653]
Reconstructing high-fidelity 3D objects from sparse, partial observation is crucial for various applications in computer vision, robotics, and graphics.
Recent neural implicit modeling methods show promising results on synthetic or dense datasets.
But, they perform poorly on real-world data that is sparse and noisy.
This paper analyzes the root cause of such deficient performance of a popular neural implicit model.
arXiv Detail & Related papers (2021-01-18T03:24:48Z) - 3D Registration for Self-Occluded Objects in Context [66.41922513553367]
We introduce the first deep learning framework capable of effectively handling this scenario.
Our method consists of an instance segmentation module followed by a pose estimation one.
It allows us to perform 3D registration in a one-shot manner, without requiring an expensive iterative procedure.
arXiv Detail & Related papers (2020-11-23T08:05:28Z) - Bridging the Reality Gap for Pose Estimation Networks using Sensor-Based
Domain Randomization [1.4290119665435117]
Methods trained on synthetic data use 2D images, as domain randomization in 2D is more developed.
Our method integrates the 3D data into the network to increase the accuracy of the pose estimation.
Experiments on three large pose estimation benchmarks show that the presented method outperforms previous methods trained on synthetic data.
arXiv Detail & Related papers (2020-11-17T09:12:11Z) - se(3)-TrackNet: Data-driven 6D Pose Tracking by Calibrating Image
Residuals in Synthetic Domains [12.71983073907091]
This work proposes a data-driven optimization approach for long-term, 6D pose tracking.
It aims to identify the optimal relative pose given the current RGB-D observation and a synthetic image conditioned on the previous best estimate and the object's model.
The proposed approach achieves consistently robust estimates and outperforms alternatives, even though they have been trained with real images.
arXiv Detail & Related papers (2020-07-27T21:09:36Z) - Cascaded deep monocular 3D human pose estimation with evolutionary
training data [76.3478675752847]
Deep representation learning has achieved remarkable accuracy for monocular 3D human pose estimation.
This paper proposes a novel data augmentation method that is scalable for massive amount of training data.
Our method synthesizes unseen 3D human skeletons based on a hierarchical human representation and synthesizings inspired by prior knowledge.
arXiv Detail & Related papers (2020-06-14T03:09:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.