MegaPose: 6D Pose Estimation of Novel Objects via Render & Compare
- URL: http://arxiv.org/abs/2212.06870v1
- Date: Tue, 13 Dec 2022 19:30:03 GMT
- Title: MegaPose: 6D Pose Estimation of Novel Objects via Render & Compare
- Authors: Yann Labb\'e, Lucas Manuelli, Arsalan Mousavian, Stephen Tyree, Stan
Birchfield, Jonathan Tremblay, Justin Carpentier, Mathieu Aubry, Dieter Fox,
Josef Sivic
- Abstract summary: MegaPose is a method to estimate the 6D pose of novel objects, that is, objects unseen during training.
We present a 6D pose refiner based on a render&compare strategy which can be applied to novel objects.
Second, we introduce a novel approach for coarse pose estimation which leverages a network trained to classify whether the pose error between a synthetic rendering and an observed image of the same object can be corrected by the refiner.
- Score: 84.80956484848505
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce MegaPose, a method to estimate the 6D pose of novel objects,
that is, objects unseen during training. At inference time, the method only
assumes knowledge of (i) a region of interest displaying the object in the
image and (ii) a CAD model of the observed object. The contributions of this
work are threefold. First, we present a 6D pose refiner based on a
render&compare strategy which can be applied to novel objects. The shape and
coordinate system of the novel object are provided as inputs to the network by
rendering multiple synthetic views of the object's CAD model. Second, we
introduce a novel approach for coarse pose estimation which leverages a network
trained to classify whether the pose error between a synthetic rendering and an
observed image of the same object can be corrected by the refiner. Third, we
introduce a large-scale synthetic dataset of photorealistic images of thousands
of objects with diverse visual and shape properties and show that this
diversity is crucial to obtain good generalization performance on novel
objects. We train our approach on this large synthetic dataset and apply it
without retraining to hundreds of novel objects in real images from several
pose estimation benchmarks. Our approach achieves state-of-the-art performance
on the ModelNet and YCB-Video datasets. An extensive evaluation on the 7 core
datasets of the BOP challenge demonstrates that our approach achieves
performance competitive with existing approaches that require access to the
target objects during training. Code, dataset and trained models are available
on the project page: https://megapose6d.github.io/.
Related papers
- FoundationPose: Unified 6D Pose Estimation and Tracking of Novel Objects [55.77542145604758]
FoundationPose is a unified foundation model for 6D object pose estimation and tracking.
Our approach can be instantly applied at test-time to a novel object without fine-tuning.
arXiv Detail & Related papers (2023-12-13T18:28:09Z) - Category-Agnostic 6D Pose Estimation with Conditional Neural Processes [19.387280883044482]
We present a novel meta-learning approach for 6D pose estimation on unknown objects.
Our algorithm learns object representation in a category-agnostic way, which endows it with strong generalization capabilities across object categories.
arXiv Detail & Related papers (2022-06-14T20:46:09Z) - FS6D: Few-Shot 6D Pose Estimation of Novel Objects [116.34922994123973]
6D object pose estimation networks are limited in their capability to scale to large numbers of object instances.
In this work, we study a new open set problem; the few-shot 6D object poses estimation: estimating the 6D pose of an unknown object by a few support views without extra training.
arXiv Detail & Related papers (2022-03-28T10:31:29Z) - OVE6D: Object Viewpoint Encoding for Depth-based 6D Object Pose
Estimation [12.773040823634908]
We propose a universal framework, called OVE6D, for model-based 6D object pose estimation from a single depth image and a target object mask.
Our model is trained using purely synthetic data rendered from ShapeNet, and, unlike most of the existing methods, it generalizes well on new real-world objects without any fine-tuning.
We show that OVE6D outperforms some contemporary deep learning-based pose estimation methods specifically trained for individual objects or datasets with real-world training data.
arXiv Detail & Related papers (2022-03-02T12:51:33Z) - DONet: Learning Category-Level 6D Object Pose and Size Estimation from
Depth Observation [53.55300278592281]
We propose a method of Category-level 6D Object Pose and Size Estimation (COPSE) from a single depth image.
Our framework makes inferences based on the rich geometric information of the object in the depth channel alone.
Our framework competes with state-of-the-art approaches that require labeled real-world images.
arXiv Detail & Related papers (2021-06-27T10:41:50Z) - Unsupervised Learning of 3D Object Categories from Videos in the Wild [75.09720013151247]
We focus on learning a model from multiple views of a large collection of object instances.
We propose a new neural network design, called warp-conditioned ray embedding (WCR), which significantly improves reconstruction.
Our evaluation demonstrates performance improvements over several deep monocular reconstruction baselines on existing benchmarks.
arXiv Detail & Related papers (2021-03-30T17:57:01Z) - Shape Prior Deformation for Categorical 6D Object Pose and Size
Estimation [62.618227434286]
We present a novel learning approach to recover the 6D poses and sizes of unseen object instances from an RGB-D image.
We propose a deep network to reconstruct the 3D object model by explicitly modeling the deformation from a pre-learned categorical shape prior.
arXiv Detail & Related papers (2020-07-16T16:45:05Z) - 3D Reconstruction of Novel Object Shapes from Single Images [23.016517962380323]
We show that our proposed SDFNet achieves state-of-the-art performance on seen and unseen shapes.
We provide the first large-scale evaluation of single image shape reconstruction to unseen objects.
arXiv Detail & Related papers (2020-06-14T00:34:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.