CrowdDriven: A New Challenging Dataset for Outdoor Visual Localization
- URL: http://arxiv.org/abs/2109.04527v1
- Date: Thu, 9 Sep 2021 19:25:48 GMT
- Title: CrowdDriven: A New Challenging Dataset for Outdoor Visual Localization
- Authors: Ara Jafarzadeh, Manuel Lopez Antequera, Pau Gargallo, Yubin Kuang,
Carl Toft, Fredrik Kahl, Torsten Sattler
- Abstract summary: We propose a new benchmark for visual localization in outdoor scenes using crowd-sourced data.
We show that our dataset is very challenging, with all evaluated methods failing on its hardest parts.
As part of the dataset release, we provide the tooling used to generate it, enabling efficient and effective 2D correspondence annotation.
- Score: 44.97567243883994
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual localization is the problem of estimating the position and orientation
from which a given image (or a sequence of images) is taken in a known scene.
It is an important part of a wide range of computer vision and robotics
applications, from self-driving cars to augmented/virtual reality systems.
Visual localization techniques should work reliably and robustly under a wide
range of conditions, including seasonal, weather, illumination and man-made
changes. Recent benchmarking efforts model this by providing images under
different conditions, and the community has made rapid progress on these
datasets since their inception. However, they are limited to a few geographical
regions and often recorded with a single device. We propose a new benchmark for
visual localization in outdoor scenes, using crowd-sourced data to cover a wide
range of geographical regions and camera devices with a focus on the failure
cases of current algorithms. Experiments with state-of-the-art localization
approaches show that our dataset is very challenging, with all evaluated
methods failing on its hardest parts. As part of the dataset release, we
provide the tooling used to generate it, enabling efficient and effective 2D
correspondence annotation to obtain reference poses.
Related papers
- Learning Where to Look: Self-supervised Viewpoint Selection for Active Localization using Geometrical Information [68.10033984296247]
This paper explores the domain of active localization, emphasizing the importance of viewpoint selection to enhance localization accuracy.
Our contributions involve using a data-driven approach with a simple architecture designed for real-time operation, a self-supervised data training method, and the capability to consistently integrate our map into a planning framework tailored for real-world robotics applications.
arXiv Detail & Related papers (2024-07-22T12:32:09Z) - GeoCLIP: Clip-Inspired Alignment between Locations and Images for
Effective Worldwide Geo-localization [61.10806364001535]
Worldwide Geo-localization aims to pinpoint the precise location of images taken anywhere on Earth.
Existing approaches divide the globe into discrete geographic cells, transforming the problem into a classification task.
We propose GeoCLIP, a novel CLIP-inspired Image-to-GPS retrieval approach that enforces alignment between the image and its corresponding GPS locations.
arXiv Detail & Related papers (2023-09-27T20:54:56Z) - 4Seasons: Benchmarking Visual SLAM and Long-Term Localization for
Autonomous Driving in Challenging Conditions [54.59279160621111]
We present a novel visual SLAM and long-term localization benchmark for autonomous driving in challenging conditions based on the large-scale 4Seasons dataset.
The proposed benchmark provides drastic appearance variations caused by seasonal changes and diverse weather and illumination conditions.
We introduce a new unified benchmark for jointly evaluating visual odometry, global place recognition, and map-based visual localization performance.
arXiv Detail & Related papers (2022-12-31T13:52:36Z) - Visual Localization via Few-Shot Scene Region Classification [84.34083435501094]
Visual (re)localization addresses the problem of estimating the 6-DoF camera pose of a query image captured in a known scene.
Recent advances in structure-based localization solve this problem by memorizing the mapping from image pixels to scene coordinates.
We propose a scene region classification approach to achieve fast and effective scene memorization with few-shot images.
arXiv Detail & Related papers (2022-08-14T22:39:02Z) - Real-time Geo-localization Using Satellite Imagery and Topography for
Unmanned Aerial Vehicles [18.71806336611299]
We propose a framework that is reliable in changing scenes and pragmatic for lightweight embedded systems on UAVs.
The framework is comprised of two stages: offline database preparation and online inference.
We present field experiments of image-based localization on two different UAV platforms to validate our results.
arXiv Detail & Related papers (2021-08-07T01:47:19Z) - Large-scale Localization Datasets in Crowded Indoor Spaces [23.071409425965772]
We introduce 5 new indoor datasets for visual localization in challenging real-world environments.
They were captured in a large shopping mall and a large metro station in Seoul, South Korea.
In order to obtain accurate ground truth camera poses, we developed a robust LiDAR SLAM.
arXiv Detail & Related papers (2021-05-19T06:20:49Z) - Self-supervised Human Detection and Segmentation via Multi-view
Consensus [116.92405645348185]
We propose a multi-camera framework in which geometric constraints are embedded in the form of multi-view consistency during training.
We show that our approach outperforms state-of-the-art self-supervised person detection and segmentation techniques on images that visually depart from those of standard benchmarks.
arXiv Detail & Related papers (2020-12-09T15:47:21Z) - Active Visual Localization in Partially Calibrated Environments [35.48595012305253]
Humans can robustly localize themselves without a map after they get lost following prominent visual cues or landmarks.
In this work, we aim at endowing autonomous agents the same ability. Such ability is important in robotics applications yet very challenging when an agent is exposed to partially calibrated environments.
We propose an indoor scene dataset ACR-6, which consists of both synthetic and real data and simulates challenging scenarios for active visual localization.
arXiv Detail & Related papers (2020-12-08T08:00:55Z) - Domain-invariant Similarity Activation Map Contrastive Learning for
Retrieval-based Long-term Visual Localization [30.203072945001136]
In this work, a general architecture is first formulated probabilistically to extract domain invariant feature through multi-domain image translation.
And then a novel gradient-weighted similarity activation mapping loss (Grad-SAM) is incorporated for finer localization with high accuracy.
Extensive experiments have been conducted to validate the effectiveness of the proposed approach on the CMUSeasons dataset.
Our performance is on par with or even outperforms the state-of-the-art image-based localization baselines in medium or high precision.
arXiv Detail & Related papers (2020-09-16T14:43:22Z) - Robust Image Retrieval-based Visual Localization using Kapture [10.249293519246478]
We present a versatile pipeline for visual localization that facilitates the use of different local and global features.
We evaluate our methods on eight public datasets where they rank top on all and first on many of them.
To foster future research, we release code, models, and all datasets used in this paper in the kapture format open source under a permissive BSD license.
arXiv Detail & Related papers (2020-07-27T21:10:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.