Appearance-Based Loop Closure Detection for Online Large-Scale and Long-Term Operation
- URL: http://arxiv.org/abs/2407.15304v1
- Date: Mon, 22 Jul 2024 00:13:00 GMT
- Title: Appearance-Based Loop Closure Detection for Online Large-Scale and Long-Term Operation
- Authors: Mathieu Labbé, François Michaud,
- Abstract summary: In appearance-based localization and mapping, loop closure detection is the process used to determinate if the current observation comes from a previously visited location or a new one.
This paper presents an online loop closure detection approach for large-scale and long-term operation.
- Score: 1.1279808969568252
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In appearance-based localization and mapping, loop closure detection is the process used to determinate if the current observation comes from a previously visited location or a new one. As the size of the internal map increases, so does the time required to compare new observations with all stored locations, eventually limiting online processing. This paper presents an online loop closure detection approach for large-scale and long-term operation. The approach is based on a memory management method, which limits the number of locations used for loop closure detection so that the computation time remains under real-time constraints. The idea consists of keeping the most recent and frequently observed locations in a Working Memory (WM) used for loop closure detection, and transferring the others into a Long-Term Memory (LTM). When a match is found between the current location and one stored in WM, associated locations stored in LTM can be updated and remembered for additional loop closure detections. Results demonstrate the approach's adaptability and scalability using ten standard data sets from other appearance-based loop closure approaches, one custom data set using real images taken over a 2 km loop of our university campus, and one custom data set (7 hours) using virtual images from the racing video game ``Need for Speed: Most Wanted''.
Related papers
- Online Temporal Action Localization with Memory-Augmented Transformer [61.39427407758131]
We propose a memory-augmented transformer (MATR) for online temporal action localization.
MATR selectively preserves the past segment features, allowing to leverage long-term context for inference.
We also propose a novel action localization method that observes the current input segment to predict the end time of the ongoing action and accesses the memory queue to estimate the start time of the action.
arXiv Detail & Related papers (2024-08-06T04:55:33Z) - Memory Management for Real-Time Appearance-Based Loop Closure Detection [1.1279808969568252]
We present a novel real-time loop closure detection approach for large-scale and long-term SLAM.
Our approach is based on a memory management method that keeps time for each new observation under a fixed limit.
arXiv Detail & Related papers (2024-07-22T00:24:12Z) - GV-Bench: Benchmarking Local Feature Matching for Geometric Verification of Long-term Loop Closure Detection [32.54601978758081]
Loop closures correct drifts in trajectory estimation to build a globally consistent map.
A false loop closure can be fatal, so verification is required as an additional step to ensure robustness.
This paper proposes a unified benchmark targeting geometric verification of loop closure detection under long-term conditional variations.
arXiv Detail & Related papers (2024-07-16T14:05:19Z) - Dual Memory Aggregation Network for Event-Based Object Detection with
Learnable Representation [79.02808071245634]
Event-based cameras are bio-inspired sensors that capture brightness change of every pixel in an asynchronous manner.
Event streams are divided into grids in the x-y-t coordinates for both positive and negative polarity, producing a set of pillars as 3D tensor representation.
Long memory is encoded in the hidden state of adaptive convLSTMs while short memory is modeled by computing spatial-temporal correlation between event pillars.
arXiv Detail & Related papers (2023-03-17T12:12:41Z) - Tracking by Associating Clips [110.08925274049409]
In this paper, we investigate an alternative by treating object association as clip-wise matching.
Our new perspective views a single long video sequence as multiple short clips, and then the tracking is performed both within and between the clips.
The benefits of this new approach are two folds. First, our method is robust to tracking error accumulation or propagation, as the video chunking allows bypassing the interrupted frames.
Second, the multiple frame information is aggregated during the clip-wise matching, resulting in a more accurate long-range track association than the current frame-wise matching.
arXiv Detail & Related papers (2022-12-20T10:33:17Z) - Loop closure detection using local 3D deep descriptors [22.93552565583209]
We present a method to address loop closure detection in simultaneous localisation and mapping using local 3D deep descriptors (L3Ds)
We propose a novel overlap measure for loop detection by computing the metric error between points that correspond to mutually-nearest-neighbour descriptors.
This novel approach enables us to accurately detect loops and estimate six degrees-of-freedom poses in the case of small overlaps.
arXiv Detail & Related papers (2021-10-31T09:18:38Z) - Temporal Memory Relation Network for Workflow Recognition from Surgical
Video [53.20825496640025]
We propose a novel end-to-end temporal memory relation network (TMNet) for relating long-range and multi-scale temporal patterns.
We have extensively validated our approach on two benchmark surgical video datasets.
arXiv Detail & Related papers (2021-03-30T13:20:26Z) - DeepVideoMVS: Multi-View Stereo on Video with Recurrent Spatio-Temporal
Fusion [67.64047158294062]
We propose an online multi-view depth prediction approach on posed video streams.
The scene geometry information computed in the previous time steps is propagated to the current time step.
We outperform the existing state-of-the-art multi-view stereo methods on most of the evaluated metrics.
arXiv Detail & Related papers (2020-12-03T18:54:03Z) - Fast and Incremental Loop Closure Detection with Deep Features and
Proximity Graphs [13.328790865796224]
This article proposes an appearance-based loop closure detection pipeline named FILD++"
The system is fed by consecutive images and, via passing them twice through a single convolutional neural network, global and local deep features are extracted.
An image-to-image pairing follows, which exploits local features to evaluate the spatial information.
arXiv Detail & Related papers (2020-09-29T02:42:47Z) - Learning Spatio-Appearance Memory Network for High-Performance Visual
Tracking [79.80401607146987]
Existing object tracking usually learns a bounding-box based template to match visual targets across frames, which cannot accurately learn a pixel-wise representation.
This paper presents a novel segmentation-based tracking architecture, which is equipped with a local-temporal memory network to learn accurate-temporal correspondence.
arXiv Detail & Related papers (2020-09-21T08:12:02Z) - Fast Loop Closure Detection via Binary Content [25.571788215801003]
We leverage and compress the information into a binary image to accelerate an existing fast loop closure detection method via binary content.
Our method is compared with the state-of-the-art loop closure detection methods and the results show that it outperforms the traditional methods at both recall rate and speed.
arXiv Detail & Related papers (2020-02-25T01:59:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.