Visual Cross-View Metric Localization with Dense Uncertainty Estimates
- URL: http://arxiv.org/abs/2208.08519v1
- Date: Wed, 17 Aug 2022 20:12:23 GMT
- Title: Visual Cross-View Metric Localization with Dense Uncertainty Estimates
- Authors: Zimin Xia, Olaf Booij, Marco Manfredi, Julian F. P. Kooij
- Abstract summary: This work addresses visual cross-view metric localization for outdoor robotics.
Given a ground-level color image and a satellite patch that contains the local surroundings, the task is to identify the location of the ground camera within the satellite patch.
We devise a novel network architecture with denser satellite descriptors, similarity matching at the bottleneck, and a dense spatial distribution as output to capture multi-modal localization ambiguities.
- Score: 11.76638109321532
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work addresses visual cross-view metric localization for outdoor
robotics. Given a ground-level color image and a satellite patch that contains
the local surroundings, the task is to identify the location of the ground
camera within the satellite patch. Related work addressed this task for
range-sensors (LiDAR, Radar), but for vision, only as a secondary regression
step after an initial cross-view image retrieval step. Since the local
satellite patch could also be retrieved through any rough localization prior
(e.g. from GPS/GNSS, temporal filtering), we drop the image retrieval objective
and focus on the metric localization only. We devise a novel network
architecture with denser satellite descriptors, similarity matching at the
bottleneck (rather than at the output as in image retrieval), and a dense
spatial distribution as output to capture multi-modal localization ambiguities.
We compare against a state-of-the-art regression baseline that uses global
image descriptors. Quantitative and qualitative experimental results on the
recently proposed VIGOR and the Oxford RobotCar datasets validate our design.
The produced probabilities are correlated with localization accuracy, and can
even be used to roughly estimate the ground camera's heading when its
orientation is unknown. Overall, our method reduces the median metric
localization error by 51%, 37%, and 28% compared to the state-of-the-art when
generalizing respectively in the same area, across areas, and across time.
Related papers
- Weakly-supervised Camera Localization by Ground-to-satellite Image Registration [52.54992898069471]
We propose a weakly supervised learning strategy for ground-to-satellite image registration.
It derives positive and negative satellite images for each ground image.
We also propose a self-supervision strategy for cross-view image relative rotation estimation.
arXiv Detail & Related papers (2024-09-10T12:57:16Z) - Boosting 3-DoF Ground-to-Satellite Camera Localization Accuracy via
Geometry-Guided Cross-View Transformer [66.82008165644892]
We propose a method to increase the accuracy of a ground camera's location and orientation by estimating the relative rotation and translation between the ground-level image and its matched/retrieved satellite image.
Experimental results demonstrate that our method significantly outperforms the state-of-the-art.
arXiv Detail & Related papers (2023-07-16T11:52:27Z) - Convolutional Cross-View Pose Estimation [9.599356978682108]
We propose a novel end-to-end method for cross-view pose estimation.
Our method is validated on the VIGOR and KITTI datasets.
On the Oxford RobotCar dataset, our method can reliably estimate the ego-vehicle's pose over time.
arXiv Detail & Related papers (2023-03-09T13:52:28Z) - Uncertainty-aware Vision-based Metric Cross-view Geolocalization [25.87104194833264]
We present an end-to-end differentiable model that uses the ground and aerial images to predict a probability distribution over possible vehicle poses.
We improve the previous state-of-the-art by a large margin even without ground or aerial data from the test region.
arXiv Detail & Related papers (2022-11-22T10:23:20Z) - CVLNet: Cross-View Semantic Correspondence Learning for Video-based
Camera Localization [89.69214577915959]
This paper tackles the problem of Cross-view Video-based camera localization.
We propose estimating the query camera's relative displacement to a satellite image before similarity matching.
Experiments have demonstrated the effectiveness of video-based localization over single image-based localization.
arXiv Detail & Related papers (2022-08-07T07:35:17Z) - Satellite Image Based Cross-view Localization for Autonomous Vehicle [59.72040418584396]
This paper shows that by using an off-the-shelf high-definition satellite image as a ready-to-use map, we are able to achieve cross-view vehicle localization up to a satisfactory accuracy.
Our method is validated on KITTI and Ford Multi-AV Seasonal datasets as ground view and Google Maps as the satellite view.
arXiv Detail & Related papers (2022-07-27T13:16:39Z) - Beyond Cross-view Image Retrieval: Highly Accurate Vehicle Localization
Using Satellite Image [91.29546868637911]
This paper addresses the problem of vehicle-mounted camera localization by matching a ground-level image with an overhead-view satellite map.
The key idea is to formulate the task as pose estimation and solve it by neural-net based optimization.
Experiments on standard autonomous vehicle localization datasets have confirmed the superiority of the proposed method.
arXiv Detail & Related papers (2022-04-10T19:16:58Z) - Accurate 3-DoF Camera Geo-Localization via Ground-to-Satellite Image
Matching [102.39635336450262]
We address the problem of ground-to-satellite image geo-localization by matching a query image captured at the ground level against a large-scale database with geotagged satellite images.
Our new method is able to achieve the fine-grained location of a query image, up to pixel size precision of the satellite image.
arXiv Detail & Related papers (2022-03-26T20:10:38Z) - City-wide Street-to-Satellite Image Geolocalization of a Mobile Ground
Agent [38.140216125792755]
Cross-view image geolocalization provides an estimate of an agent's global position by matching a local ground image to an overhead satellite image without the need for GPS.
Our approach, called Wide-Area Geolocalization (WAG), combines a neural network with a particle filter to achieve global position estimates for agents moving in GPS-denied environments.
WAG achieves position estimation accuracies on the order of 20 meters, a 98% reduction compared to a baseline training and weighting approach.
arXiv Detail & Related papers (2022-03-10T19:54:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.