DeepRING: Learning Roto-translation Invariant Representation for LiDAR
based Place Recognition
- URL: http://arxiv.org/abs/2210.11029v1
- Date: Thu, 20 Oct 2022 05:35:30 GMT
- Title: DeepRING: Learning Roto-translation Invariant Representation for LiDAR
based Place Recognition
- Authors: Sha Lu, Xuecheng Xu, Li Tang, Rong Xiong and Yue Wang
- Abstract summary: We propose DeepRING to learn the roto-translation invariant representation from LiDAR scan.
There are two keys in DeepRING: the feature is extracted from sinogram, and the feature is aggregated by magnitude spectrum.
We state the place recognition as a one-shot learning problem with each place being a class, leveraging relation learning to build representation similarity.
- Score: 12.708391665878844
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: LiDAR based place recognition is popular for loop closure detection and
re-localization. In recent years, deep learning brings improvements to place
recognition by learnable feature extraction. However, these methods degenerate
when the robot re-visits previous places with large perspective difference. To
address the challenge, we propose DeepRING to learn the roto-translation
invariant representation from LiDAR scan, so that robot visits the same place
with different perspective can have similar representations. There are two keys
in DeepRING: the feature is extracted from sinogram, and the feature is
aggregated by magnitude spectrum. The two steps keeps the final representation
with both discrimination and roto-translation invariance. Moreover, we state
the place recognition as a one-shot learning problem with each place being a
class, leveraging relation learning to build representation similarity.
Substantial experiments are carried out on public datasets, validating the
effectiveness of each proposed component, and showing that DeepRING outperforms
the comparative methods, especially in dataset level generalization.
Related papers
- RAR: Retrieving And Ranking Augmented MLLMs for Visual Recognition [78.97487780589574]
Multimodal Large Language Models (MLLMs) excel at classifying fine-grained categories.
This paper introduces a Retrieving And Ranking augmented method for MLLMs.
Our proposed approach not only addresses the inherent limitations in fine-grained recognition but also preserves the model's comprehensive knowledge base.
arXiv Detail & Related papers (2024-03-20T17:59:55Z) - Deep Homography Estimation for Visual Place Recognition [49.235432979736395]
We propose a transformer-based deep homography estimation (DHE) network.
It takes the dense feature map extracted by a backbone network as input and fits homography for fast and learnable geometric verification.
Experiments on benchmark datasets show that our method can outperform several state-of-the-art methods.
arXiv Detail & Related papers (2024-02-25T13:22:17Z) - RaSa: Relation and Sensitivity Aware Representation Learning for
Text-based Person Search [51.09723403468361]
We propose a Relation and Sensitivity aware representation learning method (RaSa)
RaSa includes two novel tasks: Relation-Aware learning (RA) and Sensitivity-Aware learning (SA)
Experiments demonstrate that RaSa outperforms existing state-of-the-art methods by 6.94%, 4.45% and 15.35% in terms of Rank@1 on datasets.
arXiv Detail & Related papers (2023-05-23T03:53:57Z) - Learning Common Rationale to Improve Self-Supervised Representation for
Fine-Grained Visual Recognition Problems [61.11799513362704]
We propose learning an additional screening mechanism to identify discriminative clues commonly seen across instances and classes.
We show that a common rationale detector can be learned by simply exploiting the GradCAM induced from the SSL objective.
arXiv Detail & Related papers (2023-03-03T02:07:40Z) - RoRD: Rotation-Robust Descriptors and Orthographic Views for Local
Feature Matching [32.10261486751993]
We present a novel framework that combines learning of invariant descriptors through data augmentation and viewpoint projection.
We evaluate the effectiveness of the proposed approach on key tasks including pose estimation and visual place recognition.
arXiv Detail & Related papers (2021-03-15T17:40:25Z) - Unsupervised Pretraining for Object Detection by Patch Reidentification [72.75287435882798]
Unsupervised representation learning achieves promising performances in pre-training representations for object detectors.
This work proposes a simple yet effective representation learning method for object detection, named patch re-identification (Re-ID)
Our method significantly outperforms its counterparts on COCO in all settings, such as different training iterations and data percentages.
arXiv Detail & Related papers (2021-03-08T15:13:59Z) - DiSCO: Differentiable Scan Context with Orientation [13.797651328615347]
We propose a LiDAR-based place recognition method, named Differentiable Scan Context with Orientation (DiSCO)
It simultaneously finds the scan at a similar place and estimates their relative orientation.
DiSCO is validated on three datasets with long-term outdoor conditions.
arXiv Detail & Related papers (2020-10-21T12:38:21Z) - Gait Recognition using Multi-Scale Partial Representation Transformation
with Capsules [22.99694601595627]
We propose a novel deep network, learning to transfer multi-scale partial gait representations using capsules.
Our network first obtains multi-scale partial representations using a state-of-the-art deep partial feature extractor.
It then recurrently learns the correlations and co-occurrences of the patterns among the partial features in forward and backward directions.
arXiv Detail & Related papers (2020-10-18T19:47:38Z) - Distilling Localization for Self-Supervised Representation Learning [82.79808902674282]
Contrastive learning has revolutionized unsupervised representation learning.
Current contrastive models are ineffective at localizing the foreground object.
We propose a data-driven approach for learning in variance to backgrounds.
arXiv Detail & Related papers (2020-04-14T16:29:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.