RING++: Roto-translation Invariant Gram for Global Localization on a
Sparse Scan Map
- URL: http://arxiv.org/abs/2210.05984v1
- Date: Wed, 12 Oct 2022 07:49:24 GMT
- Title: RING++: Roto-translation Invariant Gram for Global Localization on a
Sparse Scan Map
- Authors: Xuecheng Xu, Sha Lu, Jun Wu, Haojian Lu, Qiuguo Zhu, Yiyi Liao, Rong
Xiong and Yue Wang
- Abstract summary: We propose RING++ which has roto-translation invariant representation for place recognition, and global convergence for both rotation and translation estimation.
With the theoretical guarantee, RING++ is able to address the large viewpoint difference using a lightweight map with sparse scans.
This is the first learning-free framework to address all subtasks of global localization in the sparse scan map.
- Score: 20.276334172402763
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Global localization plays a critical role in many robot applications.
LiDAR-based global localization draws the community's focus with its robustness
against illumination and seasonal changes. To further improve the localization
under large viewpoint differences, we propose RING++ which has roto-translation
invariant representation for place recognition, and global convergence for both
rotation and translation estimation. With the theoretical guarantee, RING++ is
able to address the large viewpoint difference using a lightweight map with
sparse scans. In addition, we derive sufficient conditions of feature
extractors for the representation preserving the roto-translation invariance,
making RING++ a framework applicable to generic multi-channel features. To the
best of our knowledge, this is the first learning-free framework to address all
subtasks of global localization in the sparse scan map. Validations on
real-world datasets show that our approach demonstrates better performance than
state-of-the-art learning-free methods, and competitive performance with
learning-based methods. Finally, we integrate RING++ into a multi-robot/session
SLAM system, performing its effectiveness in collaborative applications.
Related papers
- RING#: PR-by-PE Global Localization with Roto-translation Equivariant Gram Learning [20.688641105430467]
Global localization is crucial in autonomous driving and robotics applications when GPS signals are unreliable.
Most approaches achieve global localization by sequential place recognition (PR) and pose estimation (PE)
We introduce a new paradigm, PR-by-PE localization, which bypasses the need for separate place recognition by directly deriving it from pose estimation.
We propose RING#, an end-to-end PR-by-PE localization network that operates in the bird's-eye-view (BEV) space, compatible with both vision and LiDAR sensors.
arXiv Detail & Related papers (2024-08-30T18:42:53Z) - Text-Video Retrieval with Global-Local Semantic Consistent Learning [122.15339128463715]
We propose a simple yet effective method, Global-Local Semantic Consistent Learning (GLSCL)
GLSCL capitalizes on latent shared semantics across modalities for text-video retrieval.
Our method achieves comparable performance with SOTA as well as being nearly 220 times faster in terms of computational cost.
arXiv Detail & Related papers (2024-05-21T11:59:36Z) - Unifying Local and Global Multimodal Features for Place Recognition in Aliased and Low-Texture Environments [19.859565090638167]
This paper presents a novel model, called UMF, that leverages multi-modality by cross-attention blocks between vision and LiDAR features.
Experiments, particularly on sequences captured on a planetary-analogous environment, show that UMF outperforms significantly previous baselines.
Our work aims to enhance the reliability of SLAM in all situations, we also explore its performance on the widely used RobotCar dataset.
arXiv Detail & Related papers (2024-03-20T08:35:57Z) - Adaptive Global-Local Representation Learning and Selection for
Cross-Domain Facial Expression Recognition [54.334773598942775]
Domain shift poses a significant challenge in Cross-Domain Facial Expression Recognition (CD-FER)
We propose an Adaptive Global-Local Representation Learning and Selection framework.
arXiv Detail & Related papers (2024-01-20T02:21:41Z) - GL-CLeF: A Global-Local Contrastive Learning Framework for Cross-lingual
Spoken Language Understanding [74.39024160277809]
We present Global--Local Contrastive Learning Framework (GL-CLeF) to address this shortcoming.
Specifically, we employ contrastive learning, leveraging bilingual dictionaries to construct multilingual views of the same utterance.
GL-CLeF achieves the best performance and successfully pulls representations of similar sentences across languages closer.
arXiv Detail & Related papers (2022-04-18T13:56:58Z) - Think Global, Act Local: Dual-scale Graph Transformer for
Vision-and-Language Navigation [87.03299519917019]
We propose a dual-scale graph transformer (DUET) for joint long-term action planning and fine-grained cross-modal understanding.
We build a topological map on-the-fly to enable efficient exploration in global action space.
The proposed approach, DUET, significantly outperforms state-of-the-art methods on goal-oriented vision-and-language navigation benchmarks.
arXiv Detail & Related papers (2022-02-23T19:06:53Z) - An Entropy-guided Reinforced Partial Convolutional Network for Zero-Shot
Learning [77.72330187258498]
We propose a novel Entropy-guided Reinforced Partial Convolutional Network (ERPCNet)
ERPCNet extracts and aggregates localities based on semantic relevance and visual correlations without human-annotated regions.
It not only discovers global-cooperative localities dynamically but also converges faster for policy gradient optimization.
arXiv Detail & Related papers (2021-11-03T11:13:13Z) - Global Context Aware RCNN for Object Detection [1.1939762265857436]
We propose a novel end-to-end trainable framework, called Global Context Aware (GCA) RCNN.
The core component of GCA framework is a context aware mechanism, in which both global feature pyramid and attention strategies are used for feature extraction and feature refinement.
In the end, we also present a lightweight version of our method, which only slightly increases model complexity and computational burden.
arXiv Detail & Related papers (2020-12-04T14:56:46Z) - Gait Recognition via Effective Global-Local Feature Representation and
Local Temporal Aggregation [28.721376937882958]
Gait recognition is one of the most important biometric technologies and has been applied in many fields.
Recent gait recognition frameworks represent each gait frame by descriptors extracted from either global appearances or local regions of humans.
We propose a novel feature extraction and fusion framework to achieve discriminative feature representations for gait recognition.
arXiv Detail & Related papers (2020-11-03T04:07:13Z) - Semantic Histogram Based Graph Matching for Real-Time Multi-Robot Global
Localization in Large Scale Environment [18.128244946109795]
We propose a semantic histogram-based graph matching method that is robust to viewpoint variation and can achieve real-time global localization.
Our approach is about 30 times faster than Random Walk based semantic descriptors.
It achieves an accuracy of 95% for global localization, while the accuracy of the state-of-the-art method is 85%.
arXiv Detail & Related papers (2020-10-19T08:18:42Z) - Inter-Image Communication for Weakly Supervised Localization [77.2171924626778]
Weakly supervised localization aims at finding target object regions using only image-level supervision.
We propose to leverage pixel-level similarities across different objects for learning more accurate object locations.
Our method achieves the Top-1 localization error rate of 45.17% on the ILSVRC validation set.
arXiv Detail & Related papers (2020-08-12T04:14:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.