VPR-Bench: An Open-Source Visual Place Recognition Evaluation Framework
with Quantifiable Viewpoint and Appearance Change
- URL: http://arxiv.org/abs/2005.08135v2
- Date: Fri, 1 Oct 2021 18:09:13 GMT
- Title: VPR-Bench: An Open-Source Visual Place Recognition Evaluation Framework
with Quantifiable Viewpoint and Appearance Change
- Authors: Mubariz Zaffar and Sourav Garg and Michael Milford and Julian Kooij
and David Flynn and Klaus McDonald-Maier and Shoaib Ehsan
- Abstract summary: VPR research has grown rapidly as a field over the past decade due to improving camera hardware and its potential for deep learning-based techniques.
This growth has led to fragmentation and a lack of standardisation in the field, especially concerning performance evaluation.
In this paper, we address these gaps through a new comprehensive open-source framework for assessing the performance of VPR techniques, dubbed "VPR-Bench"
- Score: 25.853640977526705
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual Place Recognition (VPR) is the process of recognising a previously
visited place using visual information, often under varying appearance
conditions and viewpoint changes and with computational constraints. VPR is
related to the concepts of localisation, loop closure, image retrieval and is a
critical component of many autonomous navigation systems ranging from
autonomous vehicles to drones and computer vision systems. While the concept of
place recognition has been around for many years, VPR research has grown
rapidly as a field over the past decade due to improving camera hardware and
its potential for deep learning-based techniques, and has become a widely
studied topic in both the computer vision and robotics communities. This growth
however has led to fragmentation and a lack of standardisation in the field,
especially concerning performance evaluation. Moreover, the notion of viewpoint
and illumination invariance of VPR techniques has largely been assessed
qualitatively and hence ambiguously in the past. In this paper, we address
these gaps through a new comprehensive open-source framework for assessing the
performance of VPR techniques, dubbed "VPR-Bench". VPR-Bench (Open-sourced at:
https://github.com/MubarizZaffar/VPR-Bench) introduces two much-needed
capabilities for VPR researchers: firstly, it contains a benchmark of 12
fully-integrated datasets and 10 VPR techniques, and secondly, it integrates a
comprehensive variation-quantified dataset for quantifying viewpoint and
illumination invariance. We apply and analyse popular evaluation metrics for
VPR from both the computer vision and robotics communities, and discuss how
these different metrics complement and/or replace each other, depending upon
the underlying applications and system requirements.
Related papers
- Breaking the Frame: Visual Place Recognition by Overlap Prediction [53.17564423756082]
We propose a novel visual place recognition approach based on overlap prediction, called VOP.
VOP proceeds co-visible image sections by obtaining patch-level embeddings using a Vision Transformer backbone.
Our approach uses a voting mechanism to assess overlap scores for potential database images.
arXiv Detail & Related papers (2024-06-23T20:00:20Z) - A Comprehensive Survey on Underwater Image Enhancement Based on Deep Learning [51.7818820745221]
Underwater image enhancement (UIE) presents a significant challenge within computer vision research.
Despite the development of numerous UIE algorithms, a thorough and systematic review is still absent.
arXiv Detail & Related papers (2024-05-30T04:46:40Z) - A-MuSIC: An Adaptive Ensemble System For Visual Place Recognition In
Changing Environments [22.58641358408613]
Visual place recognition (VPR) is an essential component of robot navigation and localization systems.
No single VPR technique excels in every environmental condition.
adaptive VPR system dubbed Adaptive Multi-Self Identification and Correction (A-MuSIC)
A-MuSIC matches or beats state-of-the-art VPR performance across all tested benchmark datasets.
arXiv Detail & Related papers (2023-03-24T19:25:22Z) - Visual Place Recognition: A Tutorial [40.576083932383895]
This paper is the first tutorial paper on visual place recognition.
It covers topics such as the formulation of the VPR problem, a general-purpose algorithmic pipeline, and an evaluation methodology for VPR approaches.
Practical code examples in Python illustrate to prospective practitioners and researchers how VPR is implemented and evaluated.
arXiv Detail & Related papers (2023-03-06T16:52:11Z) - StructVPR: Distill Structural Knowledge with Weighting Samples for
Visual Place Recognition [49.58170209388029]
Visual place recognition (VPR) is usually considered as a specific image retrieval problem.
We propose StructVPR, a novel training architecture for VPR, to enhance structural knowledge in RGB global features.
Ours achieves state-of-the-art performance while maintaining a low computational cost.
arXiv Detail & Related papers (2022-12-02T02:52:01Z) - Merging Classification Predictions with Sequential Information for
Lightweight Visual Place Recognition in Changing Environments [22.58641358408613]
Low-overhead visual place recognition (VPR) is a highly active research topic.
Mobile robotics applications often operate under low-end hardware, and even more hardware capable systems can still benefit from freeing up onboard system resources for other navigation tasks.
This work addresses lightweight VPR by proposing a novel system based on the combination of binary-weighted classifier networks with a one-dimensional convolutional network, dubbed merger.
arXiv Detail & Related papers (2022-10-03T11:42:08Z) - Delving into the Devils of Bird's-eye-view Perception: A Review,
Evaluation and Recipe [115.31507979199564]
Learning powerful representations in bird's-eye-view (BEV) for perception tasks is trending and drawing extensive attention both from industry and academia.
As sensor configurations get more complex, integrating multi-source information from different sensors and representing features in a unified view come of vital importance.
The core problems for BEV perception lie in (a) how to reconstruct the lost 3D information via view transformation from perspective view to BEV; (b) how to acquire ground truth annotations in BEV grid; and (d) how to adapt and generalize algorithms as sensor configurations vary across different scenarios.
arXiv Detail & Related papers (2022-09-12T15:29:13Z) - SwitchHit: A Probabilistic, Complementarity-Based Switching System for
Improved Visual Place Recognition in Changing Environments [20.917586014941033]
There is no universal VPR technique that can work in all types of environments.
Running multiple VPR techniques in parallel may be prohibitive for resource-constrained embedded platforms.
This paper presents a probabilistic complementarity based switching VPR system, SwitchHit.
arXiv Detail & Related papers (2022-03-01T16:23:22Z) - A Benchmark Comparison of Visual Place Recognition Techniques for
Resource-Constrained Embedded Platforms [17.48671856442762]
We present a hardware-focused benchmark evaluation of a number of state-of-the-art VPR techniques on public datasets.
We consider popular single board computers, including ODroid, UP and Raspberry Pi 3, in addition to a commodity desktop and laptop for reference.
Key questions addressed include: How does the performance accuracy of a VPR technique change with processor architecture?
The extensive analysis and results in this work serve not only as a benchmark for the VPR community, but also provide useful insights for real-world adoption of VPR applications.
arXiv Detail & Related papers (2021-09-22T19:45:57Z) - Reasoning over Vision and Language: Exploring the Benefits of
Supplemental Knowledge [59.87823082513752]
This paper investigates the injection of knowledge from general-purpose knowledge bases (KBs) into vision-and-language transformers.
We empirically study the relevance of various KBs to multiple tasks and benchmarks.
The technique is model-agnostic and can expand the applicability of any vision-and-language transformer with minimal computational overhead.
arXiv Detail & Related papers (2021-01-15T08:37:55Z) - Deep Learning for Person Re-identification: A Survey and Outlook [233.36948173686602]
Person re-identification (Re-ID) aims at retrieving a person of interest across multiple non-overlapping cameras.
By dissecting the involved components in developing a person Re-ID system, we categorize it into the closed-world and open-world settings.
arXiv Detail & Related papers (2020-01-13T12:49:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.