W-PoseNet: Dense Correspondence Regularized Pixel Pair Pose Regression
- URL: http://arxiv.org/abs/1912.11888v2
- Date: Thu, 4 Mar 2021 09:22:26 GMT
- Title: W-PoseNet: Dense Correspondence Regularized Pixel Pair Pose Regression
- Authors: Zelin Xu, Ke Chen and Kui Jia
- Abstract summary: This paper introduces a novel pose estimation algorithm W-PoseNet.
It densely regresses from input data to 6D pose and also 3D coordinates in model space.
Experiment results on the popular YCB-Video and LineMOD benchmarks show that the proposed W-PoseNet consistently achieves superior performance.
- Score: 34.8793946023412
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Solving 6D pose estimation is non-trivial to cope with intrinsic appearance
and shape variation and severe inter-object occlusion, and is made more
challenging in light of extrinsic large illumination changes and low quality of
the acquired data under an uncontrolled environment. This paper introduces a
novel pose estimation algorithm W-PoseNet, which densely regresses from input
data to 6D pose and also 3D coordinates in model space. In other words, local
features learned for pose regression in our deep network are regularized by
explicitly learning pixel-wise correspondence mapping onto 3D pose-sensitive
coordinates as an auxiliary task. Moreover, a sparse pair combination of
pixel-wise features and soft voting on pixel-pair pose predictions are designed
to improve robustness to inconsistent and sparse local features. Experiment
results on the popular YCB-Video and LineMOD benchmarks show that the proposed
W-PoseNet consistently achieves superior performance to the state-of-the-art
algorithms.
Related papers
- RelPose++: Recovering 6D Poses from Sparse-view Observations [66.6922660401558]
We address the task of estimating 6D camera poses from sparse-view image sets (2-8 images)
We build on the recent RelPose framework which learns a network that infers distributions over relative rotations over image pairs.
Our final system results in large improvements in 6D pose prediction over prior art on both seen and unseen object categories.
arXiv Detail & Related papers (2023-05-08T17:59:58Z) - CheckerPose: Progressive Dense Keypoint Localization for Object Pose
Estimation with Graph Neural Network [66.24726878647543]
Estimating the 6-DoF pose of a rigid object from a single RGB image is a crucial yet challenging task.
Recent studies have shown the great potential of dense correspondence-based solutions.
We propose a novel pose estimation algorithm named CheckerPose, which improves on three main aspects.
arXiv Detail & Related papers (2023-03-29T17:30:53Z) - Depth-based 6DoF Object Pose Estimation using Swin Transformer [1.14219428942199]
Accurately estimating the 6D pose of objects is crucial for many applications, such as robotic grasping, autonomous driving, and augmented reality.
We propose a novel framework called SwinDePose, that uses only geometric information from depth images to achieve accurate 6D pose estimation.
In experiments on the LineMod and Occlusion LineMod datasets, SwinDePose outperforms existing state-of-the-art methods for 6D object pose estimation using depth images.
arXiv Detail & Related papers (2023-03-03T18:25:07Z) - BiCo-Net: Regress Globally, Match Locally for Robust 6D Pose Estimation [32.49091033895255]
Bi-directional Correspondence Mapping Network (BiCo-Net) generates point clouds guided by a typical pose regression.
An ensemble of redundant pose predictions from locally matching and direct pose regression further refines final pose output against noisy observations.
arXiv Detail & Related papers (2022-05-07T03:37:33Z) - Coupled Iterative Refinement for 6D Multi-Object Pose Estimation [64.7198752089041]
Given a set of known 3D objects and an RGB or RGB-D input image, we detect and estimate the 6D pose of each object.
Our approach iteratively refines both pose and correspondence in a tightly coupled manner, allowing us to dynamically remove outliers to improve accuracy.
arXiv Detail & Related papers (2022-04-26T18:00:08Z) - ZebraPose: Coarse to Fine Surface Encoding for 6DoF Object Pose
Estimation [76.31125154523056]
We present a discrete descriptor, which can represent the object surface densely.
We also propose a coarse to fine training strategy, which enables fine-grained correspondence prediction.
arXiv Detail & Related papers (2022-03-17T16:16:24Z) - SO-Pose: Exploiting Self-Occlusion for Direct 6D Pose Estimation [98.83762558394345]
SO-Pose is a framework for regressing all 6 degrees-of-freedom (6DoF) for the object pose in a cluttered environment from a single RGB image.
We introduce a novel reasoning about self-occlusion, in order to establish a two-layer representation for 3D objects.
Cross-layer consistencies that align correspondences, self-occlusion and 6D pose, we can further improve accuracy and robustness.
arXiv Detail & Related papers (2021-08-18T19:49:29Z) - Uncertainty-Aware Camera Pose Estimation from Points and Lines [101.03675842534415]
Perspective-n-Point-and-Line (Pn$PL) aims at fast, accurate and robust camera localizations with respect to a 3D model from 2D-3D feature coordinates.
arXiv Detail & Related papers (2021-07-08T15:19:36Z) - 3D Point-to-Keypoint Voting Network for 6D Pose Estimation [8.801404171357916]
We propose a framework for 6D pose estimation from RGB-D data based on spatial structure characteristics of 3D keypoints.
The proposed method is verified on two benchmark datasets, LINEMOD and OCCLUSION LINEMOD.
arXiv Detail & Related papers (2020-12-22T11:43:15Z) - L6DNet: Light 6 DoF Network for Robust and Precise Object Pose
Estimation with Small Datasets [0.0]
We propose a novel approach to perform 6 DoF object pose estimation from a single RGB-D image.
We adopt a hybrid pipeline in two stages: data-driven and geometric.
Our approach is more robust and accurate than state-of-the-art methods.
arXiv Detail & Related papers (2020-02-03T17:41:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.