DCL-Net: Deep Correspondence Learning Network for 6D Pose Estimation
- URL: http://arxiv.org/abs/2210.05232v1
- Date: Tue, 11 Oct 2022 08:04:40 GMT
- Title: DCL-Net: Deep Correspondence Learning Network for 6D Pose Estimation
- Authors: Hongyang Li, Jiehong Lin and Kui Jia
- Abstract summary: We introduce a new method of Deep Correspondence Learning Network for direct 6D object pose estimation, shortened as DCL-Net.
We show that DCL-Net outperforms existing methods on three benchmarking datasets, including YCB-Video, LineMOD, and Oclussion-LineMOD.
- Score: 43.963630959349885
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Establishment of point correspondence between camera and object coordinate
systems is a promising way to solve 6D object poses. However, surrogate
objectives of correspondence learning in 3D space are a step away from the true
ones of object pose estimation, making the learning suboptimal for the end
task. In this paper, we address this shortcoming by introducing a new method of
Deep Correspondence Learning Network for direct 6D object pose estimation,
shortened as DCL-Net. Specifically, DCL-Net employs dual newly proposed Feature
Disengagement and Alignment (FDA) modules to establish, in the feature space,
partial-to-partial correspondence and complete-to-complete one for partial
object observation and its complete CAD model, respectively, which result in
aggregated pose and match feature pairs from two coordinate systems; these two
FDA modules thus bring complementary advantages. The match feature pairs are
used to learn confidence scores for measuring the qualities of deep
correspondence, while the pose feature pairs are weighted by confidence scores
for direct object pose regression. A confidence-based pose refinement network
is also proposed to further improve pose precision in an iterative manner.
Extensive experiments show that DCL-Net outperforms existing methods on three
benchmarking datasets, including YCB-Video, LineMOD, and Oclussion-LineMOD;
ablation studies also confirm the efficacy of our novel designs.
Related papers
- IPoD: Implicit Field Learning with Point Diffusion for Generalizable 3D Object Reconstruction from Single RGB-D Images [50.4538089115248]
Generalizable 3D object reconstruction from single-view RGB-D images remains a challenging task.
We propose a novel approach, IPoD, which harmonizes implicit field learning with point diffusion.
Experiments conducted on the CO3D-v2 dataset affirm the superiority of IPoD, achieving 7.8% improvement in F-score and 28.6% in Chamfer distance over existing methods.
arXiv Detail & Related papers (2024-03-30T07:17:37Z) - Q-REG: End-to-End Trainable Point Cloud Registration with Surface
Curvature [81.25511385257344]
We present a novel solution, Q-REG, which utilizes rich geometric information to estimate the rigid pose from a single correspondence.
Q-REG allows to formalize the robust estimation as an exhaustive search, hence enabling end-to-end training.
We demonstrate in the experiments that Q-REG is agnostic to the correspondence matching method and provides consistent improvement both when used only in inference and in end-to-end training.
arXiv Detail & Related papers (2023-09-27T20:58:53Z) - CheckerPose: Progressive Dense Keypoint Localization for Object Pose
Estimation with Graph Neural Network [66.24726878647543]
Estimating the 6-DoF pose of a rigid object from a single RGB image is a crucial yet challenging task.
Recent studies have shown the great potential of dense correspondence-based solutions.
We propose a novel pose estimation algorithm named CheckerPose, which improves on three main aspects.
arXiv Detail & Related papers (2023-03-29T17:30:53Z) - Category-Level 6D Object Pose and Size Estimation using Self-Supervised
Deep Prior Deformation Networks [39.6823489555449]
It is difficult to precisely annotate object instances and their semantics in 3D space, and as such, synthetic data are extensively used for these tasks.
In this work, we aim to address this issue in the task setting of Sim2Real, unsupervised domain adaptation for category-level 6D object pose and size estimation.
We propose a method that is built upon a novel CAMERA Deep Prior Deformation Network, shortened as DPDN.
arXiv Detail & Related papers (2022-07-12T10:24:52Z) - Non-Local Latent Relation Distillation for Self-Adaptive 3D Human Pose
Estimation [63.199549837604444]
3D human pose estimation approaches leverage different forms of strong (2D/3D pose) or weak (multi-view or depth) paired supervision.
We cast 3D pose learning as a self-supervised adaptation problem that aims to transfer the task knowledge from a labeled source domain to a completely unpaired target.
We evaluate different self-adaptation settings and demonstrate state-of-the-art 3D human pose estimation performance on standard benchmarks.
arXiv Detail & Related papers (2022-04-05T03:52:57Z) - DFC: Deep Feature Consistency for Robust Point Cloud Registration [0.4724825031148411]
We present a novel learning-based alignment network for complex alignment scenes.
We validate our approach on the 3DMatch dataset and the KITTI odometry dataset.
arXiv Detail & Related papers (2021-11-15T08:27:21Z) - Learning Stereopsis from Geometric Synthesis for 6D Object Pose
Estimation [11.999630902627864]
Current monocular-based 6D object pose estimation methods generally achieve less competitive results than RGBD-based methods.
This paper proposes a 3D geometric volume based pose estimation method with a short baseline two-view setting.
Experiments show that our method outperforms state-of-the-art monocular-based methods, and is robust in different objects and scenes.
arXiv Detail & Related papers (2021-09-25T02:55:05Z) - Learning 2D-3D Correspondences To Solve The Blind Perspective-n-Point
Problem [98.92148855291363]
This paper proposes a deep CNN model which simultaneously solves for both 6-DoF absolute camera pose 2D--3D correspondences.
Tests on both real and simulated data have shown that our method substantially outperforms existing approaches.
arXiv Detail & Related papers (2020-03-15T04:17:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.