Learning 2D-3D Correspondences To Solve The Blind Perspective-n-Point
Problem
- URL: http://arxiv.org/abs/2003.06752v1
- Date: Sun, 15 Mar 2020 04:17:30 GMT
- Title: Learning 2D-3D Correspondences To Solve The Blind Perspective-n-Point
Problem
- Authors: Liu Liu, Dylan Campbell, Hongdong Li, Dingfu Zhou, Xibin Song and
Ruigang Yang
- Abstract summary: This paper proposes a deep CNN model which simultaneously solves for both 6-DoF absolute camera pose 2D--3D correspondences.
Tests on both real and simulated data have shown that our method substantially outperforms existing approaches.
- Score: 98.92148855291363
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conventional absolute camera pose via a Perspective-n-Point (PnP) solver
often assumes that the correspondences between 2D image pixels and 3D points
are given. When the correspondences between 2D and 3D points are not known a
priori, the task becomes the much more challenging blind PnP problem. This
paper proposes a deep CNN model which simultaneously solves for both the 6-DoF
absolute camera pose and 2D--3D correspondences. Our model comprises three
neural modules connected in sequence. First, a two-stream PointNet-inspired
network is applied directly to both the 2D image keypoints and the 3D scene
points in order to extract discriminative point-wise features harnessing both
local and contextual information. Second, a global feature matching module is
employed to estimate a matchability matrix among all 2D--3D pairs. Third, the
obtained matchability matrix is fed into a classification module to
disambiguate inlier matches. The entire network is trained end-to-end, followed
by a robust model fitting (P3P-RANSAC) at test time only to recover the 6-DoF
camera pose. Extensive tests on both real and simulated data have shown that
our method substantially outperforms existing approaches, and is capable of
processing thousands of points a second with the state-of-the-art accuracy.
Related papers
- Robust 3D Point Clouds Classification based on Declarative Defenders [18.51700931775295]
3D point clouds are unstructured and sparse, while 2D images are structured and dense.
In this paper, we explore three distinct algorithms for mapping 3D point clouds into 2D images.
The proposed approaches demonstrate superior accuracy and robustness against adversarial attacks.
arXiv Detail & Related papers (2024-10-13T01:32:38Z) - Self-supervised Learning of LiDAR 3D Point Clouds via 2D-3D Neural Calibration [107.61458720202984]
This paper introduces a novel self-supervised learning framework for enhancing 3D perception in autonomous driving scenes.
We propose the learnable transformation alignment to bridge the domain gap between image and point cloud data.
We establish dense 2D-3D correspondences to estimate the rigid pose.
arXiv Detail & Related papers (2024-01-23T02:41:06Z) - EP2P-Loc: End-to-End 3D Point to 2D Pixel Localization for Large-Scale
Visual Localization [44.05930316729542]
We propose EP2P-Loc, a novel large-scale visual localization method for 3D point clouds.
To increase the number of inliers, we propose a simple algorithm to remove invisible 3D points in the image.
For the first time in this task, we employ a differentiable for end-to-end training.
arXiv Detail & Related papers (2023-09-14T07:06:36Z) - CheckerPose: Progressive Dense Keypoint Localization for Object Pose
Estimation with Graph Neural Network [66.24726878647543]
Estimating the 6-DoF pose of a rigid object from a single RGB image is a crucial yet challenging task.
Recent studies have shown the great potential of dense correspondence-based solutions.
We propose a novel pose estimation algorithm named CheckerPose, which improves on three main aspects.
arXiv Detail & Related papers (2023-03-29T17:30:53Z) - Partially calibrated semi-generalized pose from hybrid point
correspondences [68.22708881161049]
We study all possible camera configurations within the generalized camera system.
To derive practical solvers, we test different parameterizations as well as different solving strategies.
We show that in the presence of noise in the 3D points these solvers provide better estimates than the corresponding absolute pose solvers.
arXiv Detail & Related papers (2022-09-29T19:46:59Z) - Neural Correspondence Field for Object Pose Estimation [67.96767010122633]
We propose a method for estimating the 6DoF pose of a rigid object with an available 3D model from a single RGB image.
Unlike classical correspondence-based methods which predict 3D object coordinates at pixels of the input image, the proposed method predicts 3D object coordinates at 3D query points sampled in the camera frustum.
arXiv Detail & Related papers (2022-07-30T01:48:23Z) - OSOP: A Multi-Stage One Shot Object Pose Estimation Framework [35.89334617258322]
We present a novel one-shot method for object detection and 6 DoF pose estimation, that does not require training on target objects.
At test time, it takes as input a target image and a textured 3D query model.
We evaluate the method on LineMOD, Occlusion, Homebrewed, YCB-V and TLESS datasets.
arXiv Detail & Related papers (2022-03-29T13:12:00Z) - Weakly Supervised Learning of Keypoints for 6D Object Pose Estimation [73.40404343241782]
We propose a weakly supervised 6D object pose estimation approach based on 2D keypoint detection.
Our approach achieves comparable performance with state-of-the-art fully supervised approaches.
arXiv Detail & Related papers (2022-03-07T16:23:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.