RecNet: An Invertible Point Cloud Encoding through Range Image
Embeddings for Multi-Robot Map Sharing and Reconstruction
- URL: http://arxiv.org/abs/2402.02192v2
- Date: Mon, 4 Mar 2024 23:15:14 GMT
- Title: RecNet: An Invertible Point Cloud Encoding through Range Image
Embeddings for Multi-Robot Map Sharing and Reconstruction
- Authors: Nikolaos Stathoulopoulos, Mario A.V. Saucedo, Anton Koval and George
Nikolakopoulos
- Abstract summary: RecNet is a novel approach to effective place recognition in resource-constrained robots.
It projects 3D point clouds into range images, compresses them using an encoder-decoder framework, and subsequently reconstructs the range image, restoring the original point cloud.
Our proposed approach is assessed using both a publicly available dataset and field experiments$1$, confirming its efficacy and potential for real-world applications.
- Score: 8.602553195689513
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the field of resource-constrained robots and the need for effective place
recognition in multi-robotic systems, this article introduces RecNet, a novel
approach that concurrently addresses both challenges. The core of RecNet's
methodology involves a transformative process: it projects 3D point clouds into
range images, compresses them using an encoder-decoder framework, and
subsequently reconstructs the range image, restoring the original point cloud.
Additionally, RecNet utilizes the latent vector extracted from this process for
efficient place recognition tasks. This approach not only achieves comparable
place recognition results but also maintains a compact representation, suitable
for sharing among robots to reconstruct their collective maps. The evaluation
of RecNet encompasses an array of metrics, including place recognition
performance, the structural similarity of the reconstructed point clouds, and
the bandwidth transmission advantages, derived from sharing only the latent
vectors. Our proposed approach is assessed using both a publicly available
dataset and field experiments$^1$, confirming its efficacy and potential for
real-world applications.
Related papers
- Range and Bird's Eye View Fused Cross-Modal Visual Place Recognition [10.086473917830112]
Image-to-point cloud cross-modal Visual Place Recognition (VPR) is a challenging task where the query is an RGB image, and the database samples are LiDAR point clouds.
We propose an innovative initial retrieval + re-rank method that effectively combines information from range (or RGB) images and Bird's Eye View (BEV) images.
arXiv Detail & Related papers (2025-02-17T12:29:26Z) - LAC-Net: Linear-Fusion Attention-Guided Convolutional Network for Accurate Robotic Grasping Under the Occlusion [79.22197702626542]
This paper introduces a framework that explores amodal segmentation for robotic grasping in cluttered scenes.
We propose a Linear-fusion Attention-guided Convolutional Network (LAC-Net)
The results on different datasets show that our method achieves state-of-the-art performance.
arXiv Detail & Related papers (2024-08-06T14:50:48Z) - Boosting Cross-Domain Point Classification via Distilling Relational Priors from 2D Transformers [59.0181939916084]
Traditional 3D networks mainly focus on local geometric details and ignore the topological structure between local geometries.
We propose a novel Priors Distillation (RPD) method to extract priors from the well-trained transformers on massive images.
Experiments on the PointDA-10 and the Sim-to-Real datasets verify that the proposed method consistently achieves the state-of-the-art performance of UDA for point cloud classification.
arXiv Detail & Related papers (2024-07-26T06:29:09Z) - FRAME: A Modular Framework for Autonomous Map Merging: Advancements in the Field [12.247977717070773]
This article presents a novel approach for merging 3D point cloud maps in the context of egocentric multi-robot exploration.
The proposed approach leverages state-of-the-art place recognition and learned descriptors to efficiently detect overlap between maps.
The effectiveness of the proposed framework is successfully demonstrated through multiple field missions of robot exploration.
arXiv Detail & Related papers (2024-04-27T20:54:15Z) - Towards Efficient Object Re-Identification with A Novel Cloud-Edge Collaborative Framework [25.587445307574132]
Object re-identification (ReID) is committed to searching for objects of the same identity across cameras.
Current ReID methods assume that the deployed system follows the centralized processing paradigm.
We propose a cloud-edge collaborative framework for ReID systems, aiming to expedite the return of the desired image to the cloud server.
arXiv Detail & Related papers (2024-01-04T02:56:50Z) - Dynamic Clustering Transformer Network for Point Cloud Segmentation [23.149220817575195]
We propose a novel 3D point cloud representation network, called Dynamic Clustering Transformer Network (DCTNet)
It has an encoder-decoder architecture, allowing for both local and global feature learning.
Our method was evaluated on an object-based dataset (ShapeNet), an urban navigation dataset (Toronto-3D), and a multispectral LiDAR dataset.
arXiv Detail & Related papers (2023-05-30T01:11:05Z) - Learning Local Displacements for Point Cloud Completion [93.54286830844134]
We propose a novel approach aimed at object and semantic scene completion from a partial scan represented as a 3D point cloud.
Our architecture relies on three novel layers that are used successively within an encoder-decoder structure.
We evaluate both architectures on object and indoor scene completion tasks, achieving state-of-the-art performance.
arXiv Detail & Related papers (2022-03-30T18:31:37Z) - UPDesc: Unsupervised Point Descriptor Learning for Robust Registration [54.95201961399334]
UPDesc is an unsupervised method to learn point descriptors for robust point cloud registration.
We show that our learned descriptors yield superior performance over existing unsupervised methods.
arXiv Detail & Related papers (2021-08-05T17:11:08Z) - A Novel Upsampling and Context Convolution for Image Semantic
Segmentation [0.966840768820136]
Recent methods for semantic segmentation often employ an encoder-decoder structure using deep convolutional neural networks.
We propose a dense upsampling convolution method based on guided filtering to effectively preserve the spatial information of the image in the network.
We report a new record of 82.86% and 81.62% of pixel accuracy on ADE20K and Pascal-Context benchmark datasets, respectively.
arXiv Detail & Related papers (2021-03-20T06:16:42Z) - Unsupervised Metric Relocalization Using Transform Consistency Loss [66.19479868638925]
Training networks to perform metric relocalization traditionally requires accurate image correspondences.
We propose a self-supervised solution, which exploits a key insight: localizing a query image within a map should yield the same absolute pose, regardless of the reference image used for registration.
We evaluate our framework on synthetic and real-world data, showing our approach outperforms other supervised methods when a limited amount of ground-truth information is available.
arXiv Detail & Related papers (2020-11-01T19:24:27Z) - Weakly-Supervised Semantic Segmentation by Iterative Affinity Learning [86.45526827323954]
Weakly-supervised semantic segmentation is a challenging task as no pixel-wise label information is provided for training.
We propose an iterative algorithm to learn such pairwise relations.
We show that the proposed algorithm performs favorably against the state-of-the-art methods.
arXiv Detail & Related papers (2020-02-19T10:32:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.