Descriptellation: Deep Learned Constellation Descriptors for SLAM
- URL: http://arxiv.org/abs/2203.00567v1
- Date: Tue, 1 Mar 2022 15:43:01 GMT
- Title: Descriptellation: Deep Learned Constellation Descriptors for SLAM
- Authors: Chunwei Xing, Xinyu Sun, Andrei Cramariuc, Samuel Gull, Jen Jen Chung,
Cesar Cadena, Roland Siegwart, Florian Tschopp
- Abstract summary: We formulate a learning-based approach by constructing constellations from semantically meaningful objects.
We demonstrate the effectiveness of our Deep Learned Constellation Descriptor (Descriptellation) on the Paris-Rue-Lille and IQmulus datasets.
- Score: 33.191251654815865
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current global localization descriptors in Simultaneous Localization and
Mapping (SLAM) often fail under vast viewpoint or appearance changes. Adding
topological information of semantic objects into the descriptors ameliorates
the problem. However, hand-crafted topological descriptors extract limited
information and they are not robust to environmental noise, drastic perspective
changes, or object occlusion or misdetections. To solve this problem, we
formulate a learning-based approach by constructing constellations from
semantically meaningful objects and use Deep Graph Convolution Networks to map
the constellation representation to a descriptor. We demonstrate the
effectiveness of our Deep Learned Constellation Descriptor (Descriptellation)
on the Paris-Rue-Lille and IQmulus datasets. Although Descriptellation is
trained on randomly generated simulation datasets, it shows good generalization
abilities on real-world datasets. Descriptellation outperforms the PointNet and
handcrafted constellation descriptors for global localization, and shows
robustness against different types of noise.
Related papers
- Topology-Aware Modeling for Unsupervised Simulation-to-Reality Point Cloud Recognition [63.55828203989405]
We introduce a novel Topology-Aware Modeling (TAM) framework for Sim2Real UDA on object point clouds.<n>Our approach mitigates the domain gap by leveraging global spatial topology, characterized by low-level, high-frequency 3D structures.<n>We propose an advanced self-training strategy that combines cross-domain contrastive learning with self-training.
arXiv Detail & Related papers (2025-06-26T11:53:59Z) - CLIP-Clique: Graph-based Correspondence Matching Augmented by Vision Language Models for Object-based Global Localization [0.0]
One of the most promising approaches for localization on object maps is to use semantic graph matching.
To address the former issue, we augment the correspondence matching using Vision Language Models.
In addition, inliers are estimated deterministically using a graph-theoretic approach.
arXiv Detail & Related papers (2024-10-04T00:23:20Z) - SALSA: Swift Adaptive Lightweight Self-Attention for Enhanced LiDAR Place Recognition [9.216146804584614]
We propose SALSA, a novel, lightweight, and efficient framework for LiDAR place recognition.
It consists of a Sphereformer backbone that uses radial window attention to enable information aggregation for sparse distant points.
It outperforms existing methods on various LiDAR place recognition datasets in terms of both retrieval and metric localization while operating in real-time.
arXiv Detail & Related papers (2024-07-11T08:00:19Z) - Mesh Denoising Transformer [104.5404564075393]
Mesh denoising is aimed at removing noise from input meshes while preserving their feature structures.
SurfaceFormer is a pioneering Transformer-based mesh denoising framework.
New representation known as Local Surface Descriptor captures local geometric intricacies.
Denoising Transformer module receives the multimodal information and achieves efficient global feature aggregation.
arXiv Detail & Related papers (2024-05-10T15:27:43Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Persistent Homology Meets Object Unity: Object Recognition in Clutter [2.356908851188234]
Recognition of occluded objects in unseen and unstructured indoor environments is a challenging problem for mobile robots.
We propose a new descriptor, TOPS, for point clouds generated from depth images and an accompanying recognition framework, THOR, inspired by human reasoning.
THOR outperforms state-of-the-art methods on both the datasets and achieves substantially higher recognition accuracy for all the scenarios of the UW-IS Occluded dataset.
arXiv Detail & Related papers (2023-05-05T19:42:39Z) - Robust Change Detection Based on Neural Descriptor Fields [53.111397800478294]
We develop an object-level online change detection approach that is robust to partially overlapping observations and noisy localization results.
By associating objects via shape code similarity and comparing local object-neighbor spatial layout, our proposed approach demonstrates robustness to low observation overlap and localization noises.
arXiv Detail & Related papers (2022-08-01T17:45:36Z) - Saliency Detection via Global Context Enhanced Feature Fusion and Edge
Weighted Loss [6.112591965159383]
We propose a context fusion decoder network (CFDN) and near edge weighted loss (NEWLoss) function.
The CFDN creates an accurate saliency map by integrating global context information and thus suppressing the influence of the unnecessary spatial information.
NewLoss accelerates learning of obscure boundaries without additional modules by generating weight maps on object boundaries.
arXiv Detail & Related papers (2021-10-13T08:04:55Z) - DH3D: Deep Hierarchical 3D Descriptors for Robust Large-Scale 6DoF
Relocalization [56.15308829924527]
We propose a Siamese network that jointly learns 3D local feature detection and description directly from raw 3D points.
For detecting 3D keypoints we predict the discriminativeness of the local descriptors in an unsupervised manner.
Experiments on various benchmarks demonstrate that our method achieves competitive results for both global point cloud retrieval and local point cloud registration.
arXiv Detail & Related papers (2020-07-17T20:21:22Z) - Cross-layer Feature Pyramid Network for Salient Object Detection [102.20031050972429]
We propose a novel Cross-layer Feature Pyramid Network to improve the progressive fusion in salient object detection.
The distributed features per layer own both semantics and salient details from all other layers simultaneously, and suffer reduced loss of important information.
arXiv Detail & Related papers (2020-02-25T14:06:27Z) - SceneEncoder: Scene-Aware Semantic Segmentation of Point Clouds with A
Learnable Scene Descriptor [51.298760338410624]
We propose a SceneEncoder module to impose a scene-aware guidance to enhance the effect of global information.
The module predicts a scene descriptor, which learns to represent the categories of objects existing in the scene.
We also design a region similarity loss to propagate distinguishing features to their own neighboring points with the same label.
arXiv Detail & Related papers (2020-01-24T16:53:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.