OnePose++: Keypoint-Free One-Shot Object Pose Estimation without CAD
Models
- URL: http://arxiv.org/abs/2301.07673v1
- Date: Wed, 18 Jan 2023 17:47:13 GMT
- Title: OnePose++: Keypoint-Free One-Shot Object Pose Estimation without CAD
Models
- Authors: Xingyi He, Jiaming Sun, Yuang Wang, Di Huang, Hujun Bao, Xiaowei Zhou
- Abstract summary: OnePose relies on detecting repeatable image keypoints and is thus prone to failure on low-textured objects.
We propose a keypoint-free pose estimation pipeline to remove the need for repeatable keypoint detection.
A 2D-3D matching network directly establishes 2D-3D correspondences between the query image and the reconstructed point-cloud model.
- Score: 51.68715543630427
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a new method for object pose estimation without CAD models. The
previous feature-matching-based method OnePose has shown promising results
under a one-shot setting which eliminates the need for CAD models or
object-specific training. However, OnePose relies on detecting repeatable image
keypoints and is thus prone to failure on low-textured objects. We propose a
keypoint-free pose estimation pipeline to remove the need for repeatable
keypoint detection. Built upon the detector-free feature matching method LoFTR,
we devise a new keypoint-free SfM method to reconstruct a semi-dense
point-cloud model for the object. Given a query image for object pose
estimation, a 2D-3D matching network directly establishes 2D-3D correspondences
between the query image and the reconstructed point-cloud model without first
detecting keypoints in the image. Experiments show that the proposed pipeline
outperforms existing one-shot CAD-model-free methods by a large margin and is
comparable to CAD-model-based methods on LINEMOD even for low-textured objects.
We also collect a new dataset composed of 80 sequences of 40 low-textured
objects to facilitate future research on one-shot object pose estimation. The
supplementary material, code and dataset are available on the project page:
https://zju3dv.github.io/onepose_plus_plus/.
Related papers
- FoundationPose: Unified 6D Pose Estimation and Tracking of Novel Objects [55.77542145604758]
FoundationPose is a unified foundation model for 6D object pose estimation and tracking.
Our approach can be instantly applied at test-time to a novel object without fine-tuning.
arXiv Detail & Related papers (2023-12-13T18:28:09Z) - FoundPose: Unseen Object Pose Estimation with Foundation Features [11.32559845631345]
FoundPose is a model-based method for 6D pose estimation of unseen objects from a single RGB image.
The method can quickly onboard new objects using their 3D models without requiring any object- or task-specific training.
arXiv Detail & Related papers (2023-11-30T18:52:29Z) - GigaPose: Fast and Robust Novel Object Pose Estimation via One Correspondence [64.77224422330737]
GigaPose is a fast, robust, and accurate method for CAD-based novel object pose estimation in RGB images.
Our approach samples templates in only a two-degrees-of-freedom space instead of the usual three.
It achieves state-of-the-art accuracy and can be seamlessly integrated with existing refinement methods.
arXiv Detail & Related papers (2023-11-23T18:55:03Z) - PoseMatcher: One-shot 6D Object Pose Estimation by Deep Feature Matching [51.142988196855484]
We propose PoseMatcher, an accurate model free one-shot object pose estimator.
We create a new training pipeline for object to image matching based on a three-view system.
To enable PoseMatcher to attend to distinct input modalities, an image and a pointcloud, we introduce IO-Layer.
arXiv Detail & Related papers (2023-04-03T21:14:59Z) - NOPE: Novel Object Pose Estimation from a Single Image [67.11073133072527]
We propose an approach that takes a single image of a new object as input and predicts the relative pose of this object in new images without prior knowledge of the object's 3D model.
We achieve this by training a model to directly predict discriminative embeddings for viewpoints surrounding the object.
This prediction is done using a simple U-Net architecture with attention and conditioned on the desired pose, which yields extremely fast inference.
arXiv Detail & Related papers (2023-03-23T18:55:43Z) - OnePose: One-Shot Object Pose Estimation without CAD Models [30.307122037051126]
OnePose does not rely on CAD models and can handle objects in arbitrary categories without instance- or category-specific network training.
OnePose draws the idea from visual localization and only requires a simple RGB video scan of the object to build a sparse SfM model of the object.
To mitigate the slow runtime of existing visual localization methods, we propose a new graph attention network that directly matches 2D interest points in the query image with the 3D points in the SfM model.
arXiv Detail & Related papers (2022-05-24T17:59:21Z) - Templates for 3D Object Pose Estimation Revisited: Generalization to New
Objects and Robustness to Occlusions [79.34847067293649]
We present a method that can recognize new objects and estimate their 3D pose in RGB images even under partial occlusions.
It relies on a small set of training objects to learn local object representations.
We are the first to show generalization without retraining on the LINEMOD and Occlusion-LINEMOD datasets.
arXiv Detail & Related papers (2022-03-31T17:50:35Z) - OSOP: A Multi-Stage One Shot Object Pose Estimation Framework [35.89334617258322]
We present a novel one-shot method for object detection and 6 DoF pose estimation, that does not require training on target objects.
At test time, it takes as input a target image and a textured 3D query model.
We evaluate the method on LineMOD, Occlusion, Homebrewed, YCB-V and TLESS datasets.
arXiv Detail & Related papers (2022-03-29T13:12:00Z) - 3D Object Detection and Pose Estimation of Unseen Objects in Color
Images with Local Surface Embeddings [35.769234123059086]
We present an approach for detecting and estimating the 3D poses of objects in images that requires only an untextured CAD model.
Our approach combines Deep Learning and 3D geometry: It relies on an embedding of local 3D geometry to match the CAD models to the input images.
We show that we can use Mask-RCNN in a class-agnostic way to detect the new objects without retraining and thus drastically limit the number of possible correspondences.
arXiv Detail & Related papers (2020-10-08T15:57:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.