LocPoseNet: Robust Location Prior for Unseen Object Pose Estimation
- URL: http://arxiv.org/abs/2211.16290v3
- Date: Tue, 6 Feb 2024 14:14:23 GMT
- Title: LocPoseNet: Robust Location Prior for Unseen Object Pose Estimation
- Authors: Chen Zhao, Yinlin Hu, Mathieu Salzmann
- Abstract summary: LocPoseNet is able to robustly learn location prior for unseen objects.
Our method outperforms existing works by a large margin on LINEMOD and GenMOP.
- Score: 69.70498875887611
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Object location prior is critical for the standard 6D object pose estimation
setting. The prior can be used to initialize the 3D object translation and
facilitate 3D object rotation estimation. Unfortunately, the object detectors
that are used for this purpose do not generalize to unseen objects. Therefore,
existing 6D pose estimation methods for unseen objects either assume the
ground-truth object location to be known or yield inaccurate results when it is
unavailable. In this paper, we address this problem by developing a method,
LocPoseNet, able to robustly learn location prior for unseen objects. Our
method builds upon a template matching strategy, where we propose to distribute
the reference kernels and convolve them with a query to efficiently compute
multi-scale correlations. We then introduce a novel translation estimator,
which decouples scale-aware and scale-robust features to predict different
object location parameters. Our method outperforms existing works by a large
margin on LINEMOD and GenMOP. We further construct a challenging synthetic
dataset, which allows us to highlight the better robustness of our method to
various noise sources. Our project website is at:
https://sailor-z.github.io/projects/3DV2024_LocPoseNet.html.
Related papers
- LocaliseBot: Multi-view 3D object localisation with differentiable
rendering for robot grasping [9.690844449175948]
We focus on object pose estimation.
Our approach relies on three pieces of information: multiple views of the object, the camera's parameters at those viewpoints, and 3D CAD models of objects.
We show that the estimated object pose results in 99.65% grasp accuracy with the ground truth grasp candidates.
arXiv Detail & Related papers (2023-11-14T14:27:53Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - Coupled Iterative Refinement for 6D Multi-Object Pose Estimation [64.7198752089041]
Given a set of known 3D objects and an RGB or RGB-D input image, we detect and estimate the 6D pose of each object.
Our approach iteratively refines both pose and correspondence in a tightly coupled manner, allowing us to dynamically remove outliers to improve accuracy.
arXiv Detail & Related papers (2022-04-26T18:00:08Z) - ZebraPose: Coarse to Fine Surface Encoding for 6DoF Object Pose
Estimation [76.31125154523056]
We present a discrete descriptor, which can represent the object surface densely.
We also propose a coarse to fine training strategy, which enables fine-grained correspondence prediction.
arXiv Detail & Related papers (2022-03-17T16:16:24Z) - Objects are Different: Flexible Monocular 3D Object Detection [87.82253067302561]
We propose a flexible framework for monocular 3D object detection which explicitly decouples the truncated objects and adaptively combines multiple approaches for object depth estimation.
Experiments demonstrate that our method outperforms the state-of-the-art method by relatively 27% for the moderate level and 30% for the hard level in the test set of KITTI benchmark.
arXiv Detail & Related papers (2021-04-06T07:01:28Z) - Robust 6D Object Pose Estimation by Learning RGB-D Features [59.580366107770764]
We propose a novel discrete-continuous formulation for rotation regression to resolve this local-optimum problem.
We uniformly sample rotation anchors in SO(3), and predict a constrained deviation from each anchor to the target, as well as uncertainty scores for selecting the best prediction.
Experiments on two benchmarks: LINEMOD and YCB-Video, show that the proposed method outperforms state-of-the-art approaches.
arXiv Detail & Related papers (2020-02-29T06:24:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.