Shape Prior Deformation for Categorical 6D Object Pose and Size
Estimation
- URL: http://arxiv.org/abs/2007.08454v1
- Date: Thu, 16 Jul 2020 16:45:05 GMT
- Title: Shape Prior Deformation for Categorical 6D Object Pose and Size
Estimation
- Authors: Meng Tian, Marcelo H Ang Jr, Gim Hee Lee
- Abstract summary: We present a novel learning approach to recover the 6D poses and sizes of unseen object instances from an RGB-D image.
We propose a deep network to reconstruct the 3D object model by explicitly modeling the deformation from a pre-learned categorical shape prior.
- Score: 62.618227434286
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a novel learning approach to recover the 6D poses and sizes of
unseen object instances from an RGB-D image. To handle the intra-class shape
variation, we propose a deep network to reconstruct the 3D object model by
explicitly modeling the deformation from a pre-learned categorical shape prior.
Additionally, our network infers the dense correspondences between the depth
observation of the object instance and the reconstructed 3D model to jointly
estimate the 6D object pose and size. We design an autoencoder that trains on a
collection of object models and compute the mean latent embedding for each
category to learn the categorical shape priors. Extensive experiments on both
synthetic and real-world datasets demonstrate that our approach significantly
outperforms the state of the art. Our code is available at
https://github.com/mentian/object-deformnet.
Related papers
- GenFlow: Generalizable Recurrent Flow for 6D Pose Refinement of Novel Objects [14.598853174946656]
We present GenFlow, an approach that enables both accuracy and generalization to novel objects.
Our method predicts optical flow between the rendered image and the observed image and refines the 6D pose iteratively.
It boosts the performance by a constraint of the 3D shape and the generalizable geometric knowledge learned from an end-to-end differentiable system.
arXiv Detail & Related papers (2024-03-18T06:32:23Z) - 3D Surface Reconstruction in the Wild by Deforming Shape Priors from
Synthetic Data [24.97027425606138]
Reconstructing the underlying 3D surface of an object from a single image is a challenging problem.
We present a new method for joint category-specific 3D reconstruction and object pose estimation from a single image.
Our approach achieves state-of-the-art reconstruction performance across several real-world datasets.
arXiv Detail & Related papers (2023-02-24T20:37:27Z) - MegaPose: 6D Pose Estimation of Novel Objects via Render & Compare [84.80956484848505]
MegaPose is a method to estimate the 6D pose of novel objects, that is, objects unseen during training.
We present a 6D pose refiner based on a render&compare strategy which can be applied to novel objects.
Second, we introduce a novel approach for coarse pose estimation which leverages a network trained to classify whether the pose error between a synthetic rendering and an observed image of the same object can be corrected by the refiner.
arXiv Detail & Related papers (2022-12-13T19:30:03Z) - Category-Agnostic 6D Pose Estimation with Conditional Neural Processes [19.387280883044482]
We present a novel meta-learning approach for 6D pose estimation on unknown objects.
Our algorithm learns object representation in a category-agnostic way, which endows it with strong generalization capabilities across object categories.
arXiv Detail & Related papers (2022-06-14T20:46:09Z) - Learning Canonical 3D Object Representation for Fine-Grained Recognition [77.33501114409036]
We propose a novel framework for fine-grained object recognition that learns to recover object variation in 3D space from a single image.
We represent an object as a composition of 3D shape and its appearance, while eliminating the effect of camera viewpoint.
By incorporating 3D shape and appearance jointly in a deep representation, our method learns the discriminative representation of the object.
arXiv Detail & Related papers (2021-08-10T12:19:34Z) - DONet: Learning Category-Level 6D Object Pose and Size Estimation from
Depth Observation [53.55300278592281]
We propose a method of Category-level 6D Object Pose and Size Estimation (COPSE) from a single depth image.
Our framework makes inferences based on the rich geometric information of the object in the depth channel alone.
Our framework competes with state-of-the-art approaches that require labeled real-world images.
arXiv Detail & Related papers (2021-06-27T10:41:50Z) - From Points to Multi-Object 3D Reconstruction [71.17445805257196]
We propose a method to detect and reconstruct multiple 3D objects from a single RGB image.
A keypoint detector localizes objects as center points and directly predicts all object properties, including 9-DoF bounding boxes and 3D shapes.
The presented approach performs lightweight reconstruction in a single-stage, it is real-time capable, fully differentiable and end-to-end trainable.
arXiv Detail & Related papers (2020-12-21T18:52:21Z) - Canonical 3D Deformer Maps: Unifying parametric and non-parametric
methods for dense weakly-supervised category reconstruction [79.98689027127855]
We propose a new representation of the 3D shape of common object categories that can be learned from a collection of 2D images of independent objects.
Our method builds in a novel way on concepts from parametric deformation models, non-parametric 3D reconstruction, and canonical embeddings.
It achieves state-of-the-art results in dense 3D reconstruction on public in-the-wild datasets of faces, cars, and birds.
arXiv Detail & Related papers (2020-08-28T15:44:05Z) - Learning Canonical Shape Space for Category-Level 6D Object Pose and
Size Estimation [21.7030393344051]
We learn canonical shape space (CASS), a unified representation for a large variety of instances of a certain object category.
We train a variational auto-encoder (VAE) for generating 3D point clouds in the canonical space from an RGBD image.
VAE is trained in a cross-category fashion, exploiting the publicly available large 3D shape repositories.
arXiv Detail & Related papers (2020-01-25T14:16:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.