3D object reconstruction and 6D-pose estimation from 2D shape for
robotic grasping of objects
- URL: http://arxiv.org/abs/2203.01051v1
- Date: Wed, 2 Mar 2022 11:58:35 GMT
- Title: 3D object reconstruction and 6D-pose estimation from 2D shape for
robotic grasping of objects
- Authors: Marcell Wolnitza, Osman Kaya, Tomas Kulvicius, Florentin W\"org\"otter
and Babette Dellen
- Abstract summary: We propose a method for 3D object reconstruction and 6D-pose estimation from 2D images.
By computing transformation parameters directly from the 2D images, the number of free parameters required during the registration process is reduced.
In robot experiments, successful grasping of objects demonstrates its usability in real-world environments.
- Score: 2.330913682033217
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a method for 3D object reconstruction and 6D-pose estimation from
2D images that uses knowledge about object shape as the primary key. In the
proposed pipeline, recognition and labeling of objects in 2D images deliver 2D
segment silhouettes that are compared with the 2D silhouettes of projections
obtained from various views of a 3D model representing the recognized object
class. By computing transformation parameters directly from the 2D images, the
number of free parameters required during the registration process is reduced,
making the approach feasible. Furthermore, 3D transformations and projective
geometry are employed to arrive at a full 3D reconstruction of the object in
camera space using a calibrated set up. Inclusion of a second camera allows
resolving remaining ambiguities. The method is quantitatively evaluated using
synthetic data and tested with real data, and additional results for the
well-known Linemod data set are shown. In robot experiments, successful
grasping of objects demonstrates its usability in real-world environments, and,
where possible, a comparison with other methods is provided. The method is
applicable to scenarios where 3D object models, e.g., CAD-models or point
clouds, are available and precise pixel-wise segmentation maps of 2D images can
be obtained. Different from other methods, the method does not use 3D depth for
training, widening the domain of application.
Related papers
- Likelihood-Based Generative Radiance Field with Latent Space
Energy-Based Model for 3D-Aware Disentangled Image Representation [43.41596483002523]
We propose a likelihood-based top-down 3D-aware 2D image generative model that incorporates 3D representation via Neural Radiance Fields (NeRF) and 2D imaging process via differentiable volume rendering.
Experiments on several benchmark datasets demonstrate that the NeRF-LEBM can infer 3D object structures from 2D images, generate 2D images with novel views and objects, learn from incomplete 2D images, and learn from 2D images with known or unknown camera poses.
arXiv Detail & Related papers (2023-04-16T23:44:41Z) - SSR-2D: Semantic 3D Scene Reconstruction from 2D Images [54.46126685716471]
In this work, we explore a central 3D scene modeling task, namely, semantic scene reconstruction without using any 3D annotations.
The key idea of our approach is to design a trainable model that employs both incomplete 3D reconstructions and their corresponding source RGB-D images.
Our method achieves the state-of-the-art performance of semantic scene completion on two large-scale benchmark datasets MatterPort3D and ScanNet.
arXiv Detail & Related papers (2023-02-07T17:47:52Z) - Accelerate 3D Object Processing via Spectral Layout [1.52292571922932]
We propose to embed the essential information in a 3D object into 2D space via spectral layout.
The proposed method can achieve high quality 2D representations for 3D objects, which enables to use 2D-based methods to process 3D objects.
arXiv Detail & Related papers (2021-10-25T03:18:37Z) - AutoShape: Real-Time Shape-Aware Monocular 3D Object Detection [15.244852122106634]
We propose an approach for incorporating the shape-aware 2D/3D constraints into the 3D detection framework.
Specifically, we employ the deep neural network to learn distinguished 2D keypoints in the 2D image domain.
For generating the ground truth of 2D/3D keypoints, an automatic model-fitting approach has been proposed.
arXiv Detail & Related papers (2021-08-25T08:50:06Z) - Learning Canonical 3D Object Representation for Fine-Grained Recognition [77.33501114409036]
We propose a novel framework for fine-grained object recognition that learns to recover object variation in 3D space from a single image.
We represent an object as a composition of 3D shape and its appearance, while eliminating the effect of camera viewpoint.
By incorporating 3D shape and appearance jointly in a deep representation, our method learns the discriminative representation of the object.
arXiv Detail & Related papers (2021-08-10T12:19:34Z) - Sparse Pose Trajectory Completion [87.31270669154452]
We propose a method to learn, even using a dataset where objects appear only in sparsely sampled views.
This is achieved with a cross-modal pose trajectory transfer mechanism.
Our method is evaluated on the Pix3D and ShapeNet datasets.
arXiv Detail & Related papers (2021-05-01T00:07:21Z) - Fully Understanding Generic Objects: Modeling, Segmentation, and
Reconstruction [33.95791350070165]
Inferring 3D structure of a generic object from a 2D image is a long-standing objective of computer vision.
We take an alternative approach with semi-supervised learning. That is, for a 2D image of a generic object, we decompose it into latent representations of category, shape and albedo.
We show that the complete shape and albedo modeling enables us to leverage real 2D images in both modeling and model fitting.
arXiv Detail & Related papers (2021-04-02T02:39:29Z) - Do 2D GANs Know 3D Shape? Unsupervised 3D shape reconstruction from 2D
Image GANs [156.1209884183522]
State-of-the-art 2D generative models like GANs show unprecedented quality in modeling the natural image manifold.
We present the first attempt to directly mine 3D geometric cues from an off-the-shelf 2D GAN that is trained on RGB images only.
arXiv Detail & Related papers (2020-11-02T09:38:43Z) - Canonical 3D Deformer Maps: Unifying parametric and non-parametric
methods for dense weakly-supervised category reconstruction [79.98689027127855]
We propose a new representation of the 3D shape of common object categories that can be learned from a collection of 2D images of independent objects.
Our method builds in a novel way on concepts from parametric deformation models, non-parametric 3D reconstruction, and canonical embeddings.
It achieves state-of-the-art results in dense 3D reconstruction on public in-the-wild datasets of faces, cars, and birds.
arXiv Detail & Related papers (2020-08-28T15:44:05Z) - Cylinder3D: An Effective 3D Framework for Driving-scene LiDAR Semantic
Segmentation [87.54570024320354]
State-of-the-art methods for large-scale driving-scene LiDAR semantic segmentation often project and process the point clouds in the 2D space.
A straightforward solution to tackle the issue of 3D-to-2D projection is to keep the 3D representation and process the points in the 3D space.
We develop a 3D cylinder partition and a 3D cylinder convolution based framework, termed as Cylinder3D, which exploits the 3D topology relations and structures of driving-scene point clouds.
arXiv Detail & Related papers (2020-08-04T13:56:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.