Self-Supervised Learning of Lidar Segmentation for Autonomous Indoor
Navigation
- URL: http://arxiv.org/abs/2012.05897v1
- Date: Thu, 10 Dec 2020 18:58:10 GMT
- Title: Self-Supervised Learning of Lidar Segmentation for Autonomous Indoor
Navigation
- Authors: Hugues Thomas, Ben Agro, Mona Gridseth, Jian Zhang and Timothy D.
Barfoot
- Abstract summary: We present a self-supervised learning approach for the semantic segmentation of lidar frames.
Our method is used to train a deep point cloud segmentation architecture without any human annotation.
We provide insights into our network predictions and show that our approach can also improve the performances of common localization techniques.
- Score: 17.46116398744719
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a self-supervised learning approach for the semantic segmentation
of lidar frames. Our method is used to train a deep point cloud segmentation
architecture without any human annotation. The annotation process is automated
with the combination of simultaneous localization and mapping (SLAM) and
ray-tracing algorithms. By performing multiple navigation sessions in the same
environment, we are able to identify permanent structures, such as walls, and
disentangle short-term and long-term movable objects, such as people and
tables, respectively. New sessions can then be performed using a network
trained to predict these semantic labels. We demonstrate the ability of our
approach to improve itself over time, from one session to the next. With
semantically filtered point clouds, our robot can navigate through more complex
scenarios, which, when added to the training pool, help to improve our network
predictions. We provide insights into our network predictions and show that our
approach can also improve the performances of common localization techniques.
Related papers
- Interactive Semantic Map Representation for Skill-based Visual Object
Navigation [43.71312386938849]
This paper introduces a new representation of a scene semantic map formed during the embodied agent interaction with the indoor environment.
We have implemented this representation into a full-fledged navigation approach called SkillTron.
The proposed approach makes it possible to form both intermediate goals for robot exploration and the final goal for object navigation.
arXiv Detail & Related papers (2023-11-07T16:30:12Z) - A Spatiotemporal Correspondence Approach to Unsupervised LiDAR
Segmentation with Traffic Applications [16.260518238832887]
Key idea is to leverage the nature of a dynamic point cloud sequence and introduce drastically stronger scenarios.
We alternate between optimizing semantic into groups and clustering using point-wisetemporal labels.
Our method can learn discriminative features in an unsupervised learning fashion.
arXiv Detail & Related papers (2023-08-23T21:32:46Z) - Location-Aware Self-Supervised Transformers [74.76585889813207]
We propose to pretrain networks for semantic segmentation by predicting the relative location of image parts.
We control the difficulty of the task by masking a subset of the reference patch features visible to those of the query.
Our experiments show that this location-aware pretraining leads to representations that transfer competitively to several challenging semantic segmentation benchmarks.
arXiv Detail & Related papers (2022-12-05T16:24:29Z) - Self-supervised Pre-training for Semantic Segmentation in an Indoor
Scene [8.357801312689622]
We propose RegConsist, a method for self-supervised pre-training of a semantic segmentation model.
We use a variant of contrastive learning to train a DCNN model for predicting semantic segmentation from RGB views in the target environment.
The proposed method outperforms models pre-trained on ImageNet and achieves competitive performance when using models that are trained for exactly the same task but on a different dataset.
arXiv Detail & Related papers (2022-10-04T20:10:14Z) - Polyline Based Generative Navigable Space Segmentation for Autonomous
Visual Navigation [57.3062528453841]
We propose a representation-learning-based framework to enable robots to learn the navigable space segmentation in an unsupervised manner.
We show that the proposed PSV-Nets can learn the visual navigable space with high accuracy, even without any single label.
arXiv Detail & Related papers (2021-10-29T19:50:48Z) - X-Distill: Improving Self-Supervised Monocular Depth via Cross-Task
Distillation [69.9604394044652]
We propose a novel method to improve the self-supervised training of monocular depth via cross-task knowledge distillation.
During training, we utilize a pretrained semantic segmentation teacher network and transfer its semantic knowledge to the depth network.
We extensively evaluate the efficacy of our proposed approach on the KITTI benchmark and compare it with the latest state of the art.
arXiv Detail & Related papers (2021-10-24T19:47:14Z) - Modelling Neighbor Relation in Joint Space-Time Graph for Video
Correspondence Learning [53.74240452117145]
This paper presents a self-supervised method for learning reliable visual correspondence from unlabeled videos.
We formulate the correspondence as finding paths in a joint space-time graph, where nodes are grid patches sampled from frames, and are linked by two types of edges.
Our learned representation outperforms the state-of-the-art self-supervised methods on a variety of visual tasks.
arXiv Detail & Related papers (2021-09-28T05:40:01Z) - Understanding the Role of Individual Units in a Deep Neural Network [85.23117441162772]
We present an analytic framework to systematically identify hidden units within image classification and image generation networks.
First, we analyze a convolutional neural network (CNN) trained on scene classification and discover units that match a diverse set of object concepts.
Second, we use a similar analytic method to analyze a generative adversarial network (GAN) model trained to generate scenes.
arXiv Detail & Related papers (2020-09-10T17:59:10Z) - Self-supervised Object Tracking with Cycle-consistent Siamese Networks [55.040249900677225]
We exploit an end-to-end Siamese network in a cycle-consistent self-supervised framework for object tracking.
We propose to integrate a Siamese region proposal and mask regression network in our tracking framework so that a fast and more accurate tracker can be learned without the annotation of each frame.
arXiv Detail & Related papers (2020-08-03T04:10:38Z) - Learning Topometric Semantic Maps from Occupancy Grids [2.5234065536725963]
We propose a new approach for deriving such instance-based semantic maps purely from occupancy grids.
We employ a combination of deep learning techniques to detect, segment and extract door hypotheses from a random-sized map.
We evaluate our approach on several publicly available real-world data sets.
arXiv Detail & Related papers (2020-01-10T22:06:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.