Wide-Area Geolocalization with a Limited Field of View Camera
- URL: http://arxiv.org/abs/2209.11854v2
- Date: Thu, 18 May 2023 14:41:01 GMT
- Title: Wide-Area Geolocalization with a Limited Field of View Camera
- Authors: Lena M. Downes, Ted J. Steiner, Rebecca L. Russell, and Jonathan P.
How
- Abstract summary: Cross-view geolocalization, a supplement or replacement for GPS, localizes an agent within a search area by matching images taken from a ground-view camera to overhead images taken from satellites or aircraft.
ReWAG is a neural network and particle filter system that is able to globally localize a mobile agent in a GPS-denied environment with only odometry and a 90 degree FOV camera.
- Score: 33.34809839268686
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Cross-view geolocalization, a supplement or replacement for GPS, localizes an
agent within a search area by matching images taken from a ground-view camera
to overhead images taken from satellites or aircraft. Although the viewpoint
disparity between ground and overhead images makes cross-view geolocalization
challenging, significant progress has been made assuming that the ground agent
has access to a panoramic camera. For example, our prior work (WAG) introduced
changes in search area discretization, training loss, and particle filter
weighting that enabled city-scale panoramic cross-view geolocalization.
However, panoramic cameras are not widely used in existing robotic platforms
due to their complexity and cost. Non-panoramic cross-view geolocalization is
more applicable for robotics, but is also more challenging. This paper presents
Restricted FOV Wide-Area Geolocalization (ReWAG), a cross-view geolocalization
approach that generalizes WAG for use with standard, non-panoramic ground
cameras by creating pose-aware embeddings and providing a strategy to
incorporate particle pose into the Siamese network. ReWAG is a neural network
and particle filter system that is able to globally localize a mobile agent in
a GPS-denied environment with only odometry and a 90 degree FOV camera,
achieving similar localization accuracy as what WAG achieved with a panoramic
camera and improving localization accuracy by a factor of 100 compared to a
baseline vision transformer (ViT) approach. A video highlight that demonstrates
ReWAG's convergence on a test path of several dozen kilometers is available at
https://youtu.be/U_OBQrt8qCE.
Related papers
- Adapting Fine-Grained Cross-View Localization to Areas without Fine Ground Truth [56.565405280314884]
This paper focuses on improving the performance of a trained model in a new target area by leveraging only the target-area images without fine GT.
We propose a weakly supervised learning approach based on knowledge self-distillation.
Our approach is validated using two recent state-of-the-art models on two benchmarks.
arXiv Detail & Related papers (2024-06-01T15:58:35Z) - GeoCLIP: Clip-Inspired Alignment between Locations and Images for
Effective Worldwide Geo-localization [61.10806364001535]
Worldwide Geo-localization aims to pinpoint the precise location of images taken anywhere on Earth.
Existing approaches divide the globe into discrete geographic cells, transforming the problem into a classification task.
We propose GeoCLIP, a novel CLIP-inspired Image-to-GPS retrieval approach that enforces alignment between the image and its corresponding GPS locations.
arXiv Detail & Related papers (2023-09-27T20:54:56Z) - A Gis Aided Approach for Geolocalizing an Unmanned Aerial System Using
Deep Learning [0.4297070083645048]
We propose an alternative approach to geolocalize a UAS when GPS signal is degraded or denied.
Considering UAS has a downward-looking camera on its platform that can acquire real-time images as the platform flies, we apply modern deep learning techniques to achieve geolocalization.
We extract GIS information from OpenStreetMap (OSM) to semantically segment matched features into building and terrain classes.
arXiv Detail & Related papers (2022-08-25T17:51:15Z) - Where in the World is this Image? Transformer-based Geo-localization in
the Wild [48.69031054573838]
Predicting the geographic location (geo-localization) from a single ground-level RGB image taken anywhere in the world is a very challenging problem.
We propose TransLocator, a unified dual-branch transformer network that attends to tiny details over the entire image.
We evaluate TransLocator on four benchmark datasets - Im2GPS, Im2GPS3k, YFCC4k, YFCC26k and obtain 5.5%, 14.1%, 4.9%, 9.9% continent-level accuracy improvement.
arXiv Detail & Related papers (2022-04-29T03:27:23Z) - Beyond Cross-view Image Retrieval: Highly Accurate Vehicle Localization
Using Satellite Image [91.29546868637911]
This paper addresses the problem of vehicle-mounted camera localization by matching a ground-level image with an overhead-view satellite map.
The key idea is to formulate the task as pose estimation and solve it by neural-net based optimization.
Experiments on standard autonomous vehicle localization datasets have confirmed the superiority of the proposed method.
arXiv Detail & Related papers (2022-04-10T19:16:58Z) - Accurate 3-DoF Camera Geo-Localization via Ground-to-Satellite Image
Matching [102.39635336450262]
We address the problem of ground-to-satellite image geo-localization by matching a query image captured at the ground level against a large-scale database with geotagged satellite images.
Our new method is able to achieve the fine-grained location of a query image, up to pixel size precision of the satellite image.
arXiv Detail & Related papers (2022-03-26T20:10:38Z) - Co-visual pattern augmented generative transformer learning for
automobile geo-localization [12.449657263683337]
Cross-view geo-localization (CVGL) aims to estimate the geographical location of the ground-level camera by matching against enormous geo-tagged aerial images.
We present a novel approach using cross-view knowledge generative techniques in combination with transformers, namely mutual generative transformer learning (MGTL) for CVGL.
arXiv Detail & Related papers (2022-03-17T07:29:02Z) - City-wide Street-to-Satellite Image Geolocalization of a Mobile Ground
Agent [38.140216125792755]
Cross-view image geolocalization provides an estimate of an agent's global position by matching a local ground image to an overhead satellite image without the need for GPS.
Our approach, called Wide-Area Geolocalization (WAG), combines a neural network with a particle filter to achieve global position estimates for agents moving in GPS-denied environments.
WAG achieves position estimation accuracies on the order of 20 meters, a 98% reduction compared to a baseline training and weighting approach.
arXiv Detail & Related papers (2022-03-10T19:54:12Z) - Real-time Geo-localization Using Satellite Imagery and Topography for
Unmanned Aerial Vehicles [18.71806336611299]
We propose a framework that is reliable in changing scenes and pragmatic for lightweight embedded systems on UAVs.
The framework is comprised of two stages: offline database preparation and online inference.
We present field experiments of image-based localization on two different UAV platforms to validate our results.
arXiv Detail & Related papers (2021-08-07T01:47:19Z) - Each Part Matters: Local Patterns Facilitate Cross-view Geo-localization [54.00111565818903]
Cross-view geo-localization is to spot images of the same geographic target from different platforms.
Existing methods usually concentrate on mining the fine-grained feature of the geographic target in the image center.
We introduce a simple and effective deep neural network, called Local Pattern Network (LPN), to take advantage of contextual information.
arXiv Detail & Related papers (2020-08-26T16:06:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.