PoseMatcher: One-shot 6D Object Pose Estimation by Deep Feature Matching
- URL: http://arxiv.org/abs/2304.01382v1
- Date: Mon, 3 Apr 2023 21:14:59 GMT
- Title: PoseMatcher: One-shot 6D Object Pose Estimation by Deep Feature Matching
- Authors: Pedro Castro, Tae-Kyun Kim
- Abstract summary: We propose PoseMatcher, an accurate model free one-shot object pose estimator.
We create a new training pipeline for object to image matching based on a three-view system.
To enable PoseMatcher to attend to distinct input modalities, an image and a pointcloud, we introduce IO-Layer.
- Score: 51.142988196855484
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Estimating the pose of an unseen object is the goal of the challenging
one-shot pose estimation task. Previous methods have heavily relied on feature
matching with great success. However, these methods are often inefficient and
limited by their reliance on pre-trained models that have not be designed
specifically for pose estimation. In this paper we propose PoseMatcher, an
accurate model free one-shot object pose estimator that overcomes these
limitations. We create a new training pipeline for object to image matching
based on a three-view system: a query with a positive and negative templates.
This simple yet effective approach emulates test time scenarios by cheaply
constructing an approximation of the full object point cloud during training.
To enable PoseMatcher to attend to distinct input modalities, an image and a
pointcloud, we introduce IO-Layer, a new attention layer that efficiently
accommodates self and cross attention between the inputs. Moreover, we propose
a pruning strategy where we iteratively remove redundant regions of the target
object to further reduce the complexity and noise of the network while
maintaining accuracy. Finally we redesign commonly used pose refinement
strategies, zoom and 2D offset refinements, and adapt them to the one-shot
paradigm. We outperform all prior one-shot pose estimation methods on the
Linemod and YCB-V datasets as well achieve results rivaling recent
instance-level methods. The source code and models are available at
https://github.com/PedroCastro/PoseMatcher.
Related papers
- DVMNet: Computing Relative Pose for Unseen Objects Beyond Hypotheses [59.51874686414509]
Current approaches approximate the continuous pose representation with a large number of discrete pose hypotheses.
We present a Deep Voxel Matching Network (DVMNet) that eliminates the need for pose hypotheses and computes the relative object pose in a single pass.
Our method delivers more accurate relative pose estimates for novel objects at a lower computational cost compared to state-of-the-art methods.
arXiv Detail & Related papers (2024-03-20T15:41:32Z) - FoundationPose: Unified 6D Pose Estimation and Tracking of Novel Objects [55.77542145604758]
FoundationPose is a unified foundation model for 6D object pose estimation and tracking.
Our approach can be instantly applied at test-time to a novel object without fine-tuning.
arXiv Detail & Related papers (2023-12-13T18:28:09Z) - GS-Pose: Category-Level Object Pose Estimation via Geometric and
Semantic Correspondence [5.500735640045456]
Category-level pose estimation is a challenging task with many potential applications in computer vision and robotics.
We propose to utilize both geometric and semantic features obtained from a pre-trained foundation model.
This requires significantly less data to train than prior methods since the semantic features are robust to object texture and appearance.
arXiv Detail & Related papers (2023-11-23T02:35:38Z) - NOPE: Novel Object Pose Estimation from a Single Image [67.11073133072527]
We propose an approach that takes a single image of a new object as input and predicts the relative pose of this object in new images without prior knowledge of the object's 3D model.
We achieve this by training a model to directly predict discriminative embeddings for viewpoints surrounding the object.
This prediction is done using a simple U-Net architecture with attention and conditioned on the desired pose, which yields extremely fast inference.
arXiv Detail & Related papers (2023-03-23T18:55:43Z) - OnePose++: Keypoint-Free One-Shot Object Pose Estimation without CAD
Models [51.68715543630427]
OnePose relies on detecting repeatable image keypoints and is thus prone to failure on low-textured objects.
We propose a keypoint-free pose estimation pipeline to remove the need for repeatable keypoint detection.
A 2D-3D matching network directly establishes 2D-3D correspondences between the query image and the reconstructed point-cloud model.
arXiv Detail & Related papers (2023-01-18T17:47:13Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - ZePHyR: Zero-shot Pose Hypothesis Rating [36.52070583343388]
We introduce a novel method for zero-shot object pose estimation in clutter.
Our approach uses a hypothesis generation and scoring framework, with a focus on learning a scoring function that generalizes to objects not used for training.
We demonstrate how our system can be used by quickly scanning and building a model of a novel object, which can immediately be used by our method for pose estimation.
arXiv Detail & Related papers (2021-04-28T01:48:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.