Exploiting Multi-Layer Grid Maps for Surround-View Semantic Segmentation
of Sparse LiDAR Data
- URL: http://arxiv.org/abs/2005.06667v1
- Date: Wed, 13 May 2020 23:50:34 GMT
- Title: Exploiting Multi-Layer Grid Maps for Surround-View Semantic Segmentation
of Sparse LiDAR Data
- Authors: Frank Bieder, Sascha Wirges, Johannes Janosovits, Sven Richter,
Zheyuan Wang, and Christoph Stiller
- Abstract summary: We consider the transformation of laser range measurements into a top-view grid map representation to approach the task of LiDAR-only semantic segmentation.
We are exploiting a grid map framework to extract relevant information and represent them by using multi-layer grid maps.
We compare single-layer and multi-layer approaches and demonstrate the benefit of a multi-layer grid map input.
- Score: 2.6876976011647145
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we consider the transformation of laser range measurements
into a top-view grid map representation to approach the task of LiDAR-only
semantic segmentation. Since the recent publication of the SemanticKITTI data
set, researchers are now able to study semantic segmentation of urban LiDAR
sequences based on a reasonable amount of data. While other approaches propose
to directly learn on the 3D point clouds, we are exploiting a grid map
framework to extract relevant information and represent them by using
multi-layer grid maps. This representation allows us to use well-studied deep
learning architectures from the image domain to predict a dense semantic grid
map using only the sparse input data of a single LiDAR scan. We compare
single-layer and multi-layer approaches and demonstrate the benefit of a
multi-layer grid map input. Since the grid map representation allows us to
predict a dense, 360{\deg} semantic environment representation, we further
develop a method to combine the semantic information from multiple scans and
create dense ground truth grids. This method allows us to evaluate and compare
the performance of our models not only based on grid cells with a detection,
but on the full visible measurement range.
Related papers
- Multilayer Graph Approach to Deep Subspace Clustering [0.0]
Deep subspace clustering (DSC) networks based on self-expressive model learn representation matrix, often implemented in terms of fully connected network.
Here, we apply selected linear subspace clustering algorithm to learn representation from representations learned by all layers of encoder network including the input data.
We validate proposed approach on four well-known datasets with two DSC networks as baseline models.
arXiv Detail & Related papers (2024-01-30T14:09:41Z) - LISNeRF Mapping: LiDAR-based Implicit Mapping via Semantic Neural Fields for Large-Scale 3D Scenes [2.822816116516042]
Large-scale semantic mapping is crucial for outdoor autonomous agents to fulfill high-level tasks such as planning and navigation.
This paper proposes a novel method for large-scale 3D semantic reconstruction through implicit representations from posed LiDAR measurements alone.
arXiv Detail & Related papers (2023-11-04T03:55:38Z) - Large-Scale Multi-Hypotheses Cell Tracking Using Ultrametric Contours Maps [1.015920567871904]
We describe a method for large-scale 3D cell-tracking through a segmentation selection approach.
We show that this method achieves state-of-the-art results in 3D images from the cell tracking challenge.
Our framework is flexible and supports segmentations from off-the-shelf cell segmentation models.
arXiv Detail & Related papers (2023-08-08T18:41:38Z) - SeMLaPS: Real-time Semantic Mapping with Latent Prior Networks and
Quasi-Planar Segmentation [53.83313235792596]
We present a new methodology for real-time semantic mapping from RGB-D sequences.
It combines a 2D neural network and a 3D network based on a SLAM system with 3D occupancy mapping.
Our system achieves state-of-the-art semantic mapping quality within 2D-3D networks-based systems.
arXiv Detail & Related papers (2023-06-28T22:36:44Z) - MV-JAR: Masked Voxel Jigsaw and Reconstruction for LiDAR-Based
Self-Supervised Pre-Training [58.07391711548269]
Masked Voxel Jigsaw and Reconstruction (MV-JAR) method for LiDAR-based self-supervised pre-training.
Masked Voxel Jigsaw and Reconstruction (MV-JAR) method for LiDAR-based self-supervised pre-training.
arXiv Detail & Related papers (2023-03-23T17:59:02Z) - Rethinking Range View Representation for LiDAR Segmentation [66.73116059734788]
"Many-to-one" mapping, semantic incoherence, and shape deformation are possible impediments against effective learning from range view projections.
We present RangeFormer, a full-cycle framework comprising novel designs across network architecture, data augmentation, and post-processing.
We show that, for the first time, a range view method is able to surpass the point, voxel, and multi-view fusion counterparts in the competing LiDAR semantic and panoptic segmentation benchmarks.
arXiv Detail & Related papers (2023-03-09T16:13:27Z) - GraphCSPN: Geometry-Aware Depth Completion via Dynamic GCNs [49.55919802779889]
We propose a Graph Convolution based Spatial Propagation Network (GraphCSPN) as a general approach for depth completion.
In this work, we leverage convolution neural networks as well as graph neural networks in a complementary way for geometric representation learning.
Our method achieves the state-of-the-art performance, especially when compared in the case of using only a few propagation steps.
arXiv Detail & Related papers (2022-10-19T17:56:03Z) - Improving Lidar-Based Semantic Segmentation of Top-View Grid Maps by
Learning Features in Complementary Representations [3.0413873719021995]
We introduce a novel way to predict semantic information from sparse, single-shot LiDAR measurements in the context of autonomous driving.
The approach is aimed specifically at improving the semantic segmentation of top-view grid maps.
For each representation a tailored deep learning architecture is developed to effectively extract semantic information.
arXiv Detail & Related papers (2022-03-02T14:49:51Z) - TANDEM: Tracking and Dense Mapping in Real-time using Deep Multi-view
Stereo [55.30992853477754]
We present TANDEM, a real-time monocular tracking and dense framework.
For pose estimation, TANDEM performs photometric bundle adjustment based on a sliding window of alignments.
TANDEM shows state-of-the-art real-time 3D reconstruction performance.
arXiv Detail & Related papers (2021-11-14T19:01:02Z) - PillarSegNet: Pillar-based Semantic Grid Map Estimation using Sparse
LiDAR Data [2.4038276485632846]
We propose PillarSegNet to be able to output a dense semantic grid map.
In contrast to a previously proposed grid map method, PillarSegNet uses PointNet to learn features directly from the 3D point cloud.
We show that PillarSegNet achieves a performance gain of about 10% mIoU over the state-of-the-art grid map method.
arXiv Detail & Related papers (2021-05-10T08:03:11Z) - Campus3D: A Photogrammetry Point Cloud Benchmark for Hierarchical
Understanding of Outdoor Scene [76.4183572058063]
We present a richly-annotated 3D point cloud dataset for multiple outdoor scene understanding tasks.
The dataset has been point-wisely annotated with both hierarchical and instance-based labels.
We formulate a hierarchical learning problem for 3D point cloud segmentation and propose a measurement evaluating consistency across various hierarchies.
arXiv Detail & Related papers (2020-08-11T19:10:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.