DGECN: A Depth-Guided Edge Convolutional Network for End-to-End 6D Pose
Estimation
- URL: http://arxiv.org/abs/2204.09983v1
- Date: Thu, 21 Apr 2022 09:19:50 GMT
- Title: DGECN: A Depth-Guided Edge Convolutional Network for End-to-End 6D Pose
Estimation
- Authors: Tuo Cao, Fei Luo, Yanping Fu, Wenxiao Zhang, Shengjie Zheng, Chunxia
Xiao
- Abstract summary: We propose a Depth-Guided Edge Conal Network (DGECN) for 6D pose estimation task.
We take advantages ofestimated depth information to guide both the correspondences-extraction process and the cascaded differentiable RANSAC algorithm with geometric information.
Experiments demonstrate that our proposed network outperforms current works on both effectiveness and efficiency.
- Score: 19.303780745324502
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Monocular 6D pose estimation is a fundamental task in computer vision.
Existing works often adopt a two-stage pipeline by establishing correspondences
and utilizing a RANSAC algorithm to calculate 6 degrees-of-freedom (6DoF) pose.
Recent works try to integrate differentiable RANSAC algorithms to achieve an
end-to-end 6D pose estimation. However, most of them hardly consider the
geometric features in 3D space, and ignore the topology cues when performing
differentiable RANSAC algorithms. To this end, we proposed a Depth-Guided Edge
Convolutional Network (DGECN) for 6D pose estimation task. We have made efforts
from the following three aspects: 1) We take advantages ofestimated depth
information to guide both the correspondences-extraction process and the
cascaded differentiable RANSAC algorithm with geometric information. 2)We
leverage the uncertainty ofthe estimated depth map to improve accuracy and
robustness ofthe output 6D pose. 3) We propose a differentiable
Perspective-n-Point(PnP) algorithm via edge convolution to explore the topology
relations between 2D-3D correspondences. Experiments demonstrate that our
proposed network outperforms current works on both effectiveness and
efficiency.
Related papers
- Occupancy-Based Dual Contouring [12.944046673902415]
We introduce a dual contouring method that provides state-of-the-art performance for occupancy functions.
Our method is learning-free and carefully designed to maximize the use of GPU parallelization.
arXiv Detail & Related papers (2024-09-20T11:32:21Z) - Spatial Feature Mapping for 6DoF Object Pose Estimation [29.929911622127502]
This work aims to estimate 6Dof (6D) object pose in background clutter.
Considering the strong occlusion and background noise, we propose to utilize the spatial structure for better tackling this challenging task.
arXiv Detail & Related papers (2022-06-03T21:44:10Z) - Coupled Iterative Refinement for 6D Multi-Object Pose Estimation [64.7198752089041]
Given a set of known 3D objects and an RGB or RGB-D input image, we detect and estimate the 6D pose of each object.
Our approach iteratively refines both pose and correspondence in a tightly coupled manner, allowing us to dynamically remove outliers to improve accuracy.
arXiv Detail & Related papers (2022-04-26T18:00:08Z) - Weakly Supervised Learning of Keypoints for 6D Object Pose Estimation [73.40404343241782]
We propose a weakly supervised 6D object pose estimation approach based on 2D keypoint detection.
Our approach achieves comparable performance with state-of-the-art fully supervised approaches.
arXiv Detail & Related papers (2022-03-07T16:23:47Z) - Learning Stereopsis from Geometric Synthesis for 6D Object Pose
Estimation [11.999630902627864]
Current monocular-based 6D object pose estimation methods generally achieve less competitive results than RGBD-based methods.
This paper proposes a 3D geometric volume based pose estimation method with a short baseline two-view setting.
Experiments show that our method outperforms state-of-the-art monocular-based methods, and is robust in different objects and scenes.
arXiv Detail & Related papers (2021-09-25T02:55:05Z) - SO-Pose: Exploiting Self-Occlusion for Direct 6D Pose Estimation [98.83762558394345]
SO-Pose is a framework for regressing all 6 degrees-of-freedom (6DoF) for the object pose in a cluttered environment from a single RGB image.
We introduce a novel reasoning about self-occlusion, in order to establish a two-layer representation for 3D objects.
Cross-layer consistencies that align correspondences, self-occlusion and 6D pose, we can further improve accuracy and robustness.
arXiv Detail & Related papers (2021-08-18T19:49:29Z) - FS-Net: Fast Shape-based Network for Category-Level 6D Object Pose
Estimation with Decoupled Rotation Mechanism [49.89268018642999]
We propose a fast shape-based network (FS-Net) with efficient category-level feature extraction for 6D pose estimation.
The proposed method achieves state-of-the-art performance in both category- and instance-level 6D object pose estimation.
arXiv Detail & Related papers (2021-03-12T03:07:24Z) - GDR-Net: Geometry-Guided Direct Regression Network for Monocular 6D
Object Pose Estimation [71.83992173720311]
6D pose estimation from a single RGB image is a fundamental task in computer vision.
We propose a simple yet effective Geometry-guided Direct Regression Network (GDR-Net) to learn the 6D pose in an end-to-end manner.
Our approach remarkably outperforms state-of-the-art methods on LM, LM-O and YCB-V datasets.
arXiv Detail & Related papers (2021-02-24T09:11:31Z) - Learning 2D-3D Correspondences To Solve The Blind Perspective-n-Point
Problem [98.92148855291363]
This paper proposes a deep CNN model which simultaneously solves for both 6-DoF absolute camera pose 2D--3D correspondences.
Tests on both real and simulated data have shown that our method substantially outperforms existing approaches.
arXiv Detail & Related papers (2020-03-15T04:17:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.