BiCo-Net: Regress Globally, Match Locally for Robust 6D Pose Estimation
- URL: http://arxiv.org/abs/2205.03536v1
- Date: Sat, 7 May 2022 03:37:33 GMT
- Title: BiCo-Net: Regress Globally, Match Locally for Robust 6D Pose Estimation
- Authors: Zelin Xu, Yichen Zhang, Ke Chen, Kui Jia
- Abstract summary: Bi-directional Correspondence Mapping Network (BiCo-Net) generates point clouds guided by a typical pose regression.
An ensemble of redundant pose predictions from locally matching and direct pose regression further refines final pose output against noisy observations.
- Score: 32.49091033895255
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The challenges of learning a robust 6D pose function lie in 1) severe
occlusion and 2) systematic noises in depth images. Inspired by the success of
point-pair features, the goal of this paper is to recover the 6D pose of an
object instance segmented from RGB-D images by locally matching pairs of
oriented points between the model and camera space. To this end, we propose a
novel Bi-directional Correspondence Mapping Network (BiCo-Net) to first
generate point clouds guided by a typical pose regression, which can thus
incorporate pose-sensitive information to optimize generation of local
coordinates and their normal vectors. As pose predictions via geometric
computation only rely on one single pair of local oriented points, our BiCo-Net
can achieve robustness against sparse and occluded point clouds. An ensemble of
redundant pose predictions from locally matching and direct pose regression
further refines final pose output against noisy observations. Experimental
results on three popularly benchmarking datasets can verify that our method can
achieve state-of-the-art performance, especially for the more challenging
severe occluded scenes. Source codes are available at
https://github.com/Gorilla-Lab-SCUT/BiCo-Net.
Related papers
- RDPN6D: Residual-based Dense Point-wise Network for 6Dof Object Pose Estimation Based on RGB-D Images [13.051302134031808]
We introduce a novel method for calculating the 6DoF pose of an object using a single RGB-D image.
Unlike existing methods that either directly predict objects' poses or rely on sparse keypoints for pose recovery, our approach addresses this challenging task using dense correspondence.
arXiv Detail & Related papers (2024-05-14T10:10:45Z) - RGB-based Category-level Object Pose Estimation via Decoupled Metric
Scale Recovery [72.13154206106259]
We propose a novel pipeline that decouples the 6D pose and size estimation to mitigate the influence of imperfect scales on rigid transformations.
Specifically, we leverage a pre-trained monocular estimator to extract local geometric information.
A separate branch is designed to directly recover the metric scale of the object based on category-level statistics.
arXiv Detail & Related papers (2023-09-19T02:20:26Z) - RelPose++: Recovering 6D Poses from Sparse-view Observations [66.6922660401558]
We address the task of estimating 6D camera poses from sparse-view image sets (2-8 images)
We build on the recent RelPose framework which learns a network that infers distributions over relative rotations over image pairs.
Our final system results in large improvements in 6D pose prediction over prior art on both seen and unseen object categories.
arXiv Detail & Related papers (2023-05-08T17:59:58Z) - CheckerPose: Progressive Dense Keypoint Localization for Object Pose
Estimation with Graph Neural Network [66.24726878647543]
Estimating the 6-DoF pose of a rigid object from a single RGB image is a crucial yet challenging task.
Recent studies have shown the great potential of dense correspondence-based solutions.
We propose a novel pose estimation algorithm named CheckerPose, which improves on three main aspects.
arXiv Detail & Related papers (2023-03-29T17:30:53Z) - A Dynamic Keypoints Selection Network for 6DoF Pose Estimation [0.0]
6 DoF poses estimation problem aims to estimate the rotation and translation parameters between two coordinates.
We present a novel deep neural network based on dynamic keypoints selection designed for 6DoF pose estimation from a single RGBD image.
arXiv Detail & Related papers (2021-10-24T09:58:56Z) - SO-Pose: Exploiting Self-Occlusion for Direct 6D Pose Estimation [98.83762558394345]
SO-Pose is a framework for regressing all 6 degrees-of-freedom (6DoF) for the object pose in a cluttered environment from a single RGB image.
We introduce a novel reasoning about self-occlusion, in order to establish a two-layer representation for 3D objects.
Cross-layer consistencies that align correspondences, self-occlusion and 6D pose, we can further improve accuracy and robustness.
arXiv Detail & Related papers (2021-08-18T19:49:29Z) - CAPTRA: CAtegory-level Pose Tracking for Rigid and Articulated Objects
from Point Clouds [97.63549045541296]
We propose a unified framework that can handle 9DoF pose tracking for novel rigid object instances and per-part pose tracking for articulated objects.
Our method achieves new state-of-the-art performance on category-level rigid object pose (NOCS-REAL275) and articulated object pose benchmarks (SAPIEN, BMVC) at the fastest FPS 12.
arXiv Detail & Related papers (2021-04-08T00:14:58Z) - Deep Bingham Networks: Dealing with Uncertainty and Ambiguity in Pose
Estimation [74.76155168705975]
Deep Bingham Networks (DBN) can handle pose-related uncertainties and ambiguities arising in almost all real life applications concerning 3D data.
DBN extends the state of the art direct pose regression networks by (i) a multi-hypotheses prediction head which can yield different distribution modes.
We propose new training strategies so as to avoid mode or posterior collapse during training and to improve numerical stability.
arXiv Detail & Related papers (2020-12-20T19:20:26Z) - W-PoseNet: Dense Correspondence Regularized Pixel Pair Pose Regression [34.8793946023412]
This paper introduces a novel pose estimation algorithm W-PoseNet.
It densely regresses from input data to 6D pose and also 3D coordinates in model space.
Experiment results on the popular YCB-Video and LineMOD benchmarks show that the proposed W-PoseNet consistently achieves superior performance.
arXiv Detail & Related papers (2019-12-26T15:51:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.