CyberLoc: Towards Accurate Long-term Visual Localization
- URL: http://arxiv.org/abs/2301.02403v1
- Date: Fri, 6 Jan 2023 06:49:36 GMT
- Title: CyberLoc: Towards Accurate Long-term Visual Localization
- Authors: Liu Liu, Yukai Lin, Xiao Liang, Qichao Xu, Miao Jia, Yangdong Liu,
Yuxiang Wen, Wei Luo, Jiangwei Li
- Abstract summary: CyberLoc is an image-based visual localization pipeline for robust and accurate long-term estimation under challenging conditions.
A mapping module is applied to build accurate 3D maps of a scene, one map for each reference sequence if there exist multiple sequences under different conditions.
A single-image-based localization pipeline is performed to estimate 6-DoF camera poses for each query image, one for each 3D map.
- Score: 14.883028430847714
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This technical report introduces CyberLoc, an image-based visual localization
pipeline for robust and accurate long-term pose estimation under challenging
conditions. The proposed method comprises four modules connected in a sequence.
First, a mapping module is applied to build accurate 3D maps of the scene, one
map for each reference sequence if there exist multiple reference sequences
under different conditions. Second, a single-image-based localization pipeline
(retrieval--matching--PnP) is performed to estimate 6-DoF camera poses for each
query image, one for each 3D map. Third, a consensus set maximization module is
proposed to filter out outlier 6-DoF camera poses, and outputs one 6-DoF camera
pose for a query. Finally, a robust pose refinement module is proposed to
optimize 6-DoF query poses, taking candidate global 6-DoF camera poses and
their corresponding global 2D-3D matches, sparse 2D-2D feature matches between
consecutive query images and SLAM poses of the query sequence as input.
Experiments on the 4seasons dataset show that our method achieves high accuracy
and robustness. In particular, our approach wins the localization challenge of
ECCV 2022 workshop on Map-based Localization for Autonomous Driving
(MLAD-ECCV2022).
Related papers
- Learning to Estimate 6DoF Pose from Limited Data: A Few-Shot,
Generalizable Approach using RGB Images [60.0898989456276]
We present a new framework named Cas6D for few-shot 6DoF pose estimation that is generalizable and uses only RGB images.
To address the false positives of target object detection in the extreme few-shot setting, our framework utilizes a self-supervised pre-trained ViT to learn robust feature representations.
Experimental results on the LINEMOD and GenMOP datasets demonstrate that Cas6D outperforms state-of-the-art methods by 9.2% and 3.8% accuracy (Proj-5) under the 32-shot setting.
arXiv Detail & Related papers (2023-06-13T07:45:42Z) - DPODv2: Dense Correspondence-Based 6 DoF Pose Estimation [24.770767430749288]
We propose a 3 stage 6 DoF object detection method called DPODv2 (Dense Pose Object Detector)
We combine a 2D object detector with a dense correspondence estimation network and a multi-view pose refinement method to estimate a full 6 DoF pose.
DPODv2 achieves excellent results on all of them while still remaining fast and scalable independent of the used data modality and the type of training data.
arXiv Detail & Related papers (2022-07-06T16:48:56Z) - BiCo-Net: Regress Globally, Match Locally for Robust 6D Pose Estimation [32.49091033895255]
Bi-directional Correspondence Mapping Network (BiCo-Net) generates point clouds guided by a typical pose regression.
An ensemble of redundant pose predictions from locally matching and direct pose regression further refines final pose output against noisy observations.
arXiv Detail & Related papers (2022-05-07T03:37:33Z) - Coupled Iterative Refinement for 6D Multi-Object Pose Estimation [64.7198752089041]
Given a set of known 3D objects and an RGB or RGB-D input image, we detect and estimate the 6D pose of each object.
Our approach iteratively refines both pose and correspondence in a tightly coupled manner, allowing us to dynamically remove outliers to improve accuracy.
arXiv Detail & Related papers (2022-04-26T18:00:08Z) - SO-Pose: Exploiting Self-Occlusion for Direct 6D Pose Estimation [98.83762558394345]
SO-Pose is a framework for regressing all 6 degrees-of-freedom (6DoF) for the object pose in a cluttered environment from a single RGB image.
We introduce a novel reasoning about self-occlusion, in order to establish a two-layer representation for 3D objects.
Cross-layer consistencies that align correspondences, self-occlusion and 6D pose, we can further improve accuracy and robustness.
arXiv Detail & Related papers (2021-08-18T19:49:29Z) - CAPTRA: CAtegory-level Pose Tracking for Rigid and Articulated Objects
from Point Clouds [97.63549045541296]
We propose a unified framework that can handle 9DoF pose tracking for novel rigid object instances and per-part pose tracking for articulated objects.
Our method achieves new state-of-the-art performance on category-level rigid object pose (NOCS-REAL275) and articulated object pose benchmarks (SAPIEN, BMVC) at the fastest FPS 12.
arXiv Detail & Related papers (2021-04-08T00:14:58Z) - Multi-View Multi-Person 3D Pose Estimation with Plane Sweep Stereo [71.59494156155309]
Existing approaches for multi-view 3D pose estimation explicitly establish cross-view correspondences to group 2D pose detections from multiple camera views.
We present our multi-view 3D pose estimation approach based on plane sweep stereo to jointly address the cross-view fusion and 3D pose reconstruction in a single shot.
arXiv Detail & Related papers (2021-04-06T03:49:35Z) - Learning 2D-3D Correspondences To Solve The Blind Perspective-n-Point
Problem [98.92148855291363]
This paper proposes a deep CNN model which simultaneously solves for both 6-DoF absolute camera pose 2D--3D correspondences.
Tests on both real and simulated data have shown that our method substantially outperforms existing approaches.
arXiv Detail & Related papers (2020-03-15T04:17:30Z) - Deep-Geometric 6 DoF Localization from a Single Image in Topo-metric
Maps [39.05304338751328]
We describe a Deep-Geometric Localizer that is able to estimate the full 6 Degree of Freedom (DoF) global pose of the camera from a single image.
Our method divorces the mapping and the localization algorithms (stereo and mono) and allows accurate 6 DoF pose estimation in a previously mapped environment.
With potential VR/AR and localization applications in single camera devices such as mobile phones and drones, our hybrid algorithm compares favourably with the fully Deep-Learning based Pose-Net.
arXiv Detail & Related papers (2020-02-04T10:11:46Z) - L6DNet: Light 6 DoF Network for Robust and Precise Object Pose
Estimation with Small Datasets [0.0]
We propose a novel approach to perform 6 DoF object pose estimation from a single RGB-D image.
We adopt a hybrid pipeline in two stages: data-driven and geometric.
Our approach is more robust and accurate than state-of-the-art methods.
arXiv Detail & Related papers (2020-02-03T17:41:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.