PillarSegNet: Pillar-based Semantic Grid Map Estimation using Sparse
LiDAR Data
- URL: http://arxiv.org/abs/2105.04169v1
- Date: Mon, 10 May 2021 08:03:11 GMT
- Title: PillarSegNet: Pillar-based Semantic Grid Map Estimation using Sparse
LiDAR Data
- Authors: Juncong Fei, Kunyu Peng, Philipp Heidenreich, Frank Bieder and
Christoph Stiller
- Abstract summary: We propose PillarSegNet to be able to output a dense semantic grid map.
In contrast to a previously proposed grid map method, PillarSegNet uses PointNet to learn features directly from the 3D point cloud.
We show that PillarSegNet achieves a performance gain of about 10% mIoU over the state-of-the-art grid map method.
- Score: 2.4038276485632846
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Semantic understanding of the surrounding environment is essential for
automated vehicles. The recent publication of the SemanticKITTI dataset
stimulates the research on semantic segmentation of LiDAR point clouds in urban
scenarios. While most existing approaches predict sparse pointwise semantic
classes for the sparse input LiDAR scan, we propose PillarSegNet to be able to
output a dense semantic grid map. In contrast to a previously proposed grid map
method, PillarSegNet uses PointNet to learn features directly from the 3D point
cloud and then conducts 2D semantic segmentation in the top view. To train and
evaluate our approach, we use both sparse and dense ground truth, where the
dense ground truth is obtained from multiple superimposed scans. Experimental
results on the SemanticKITTI dataset show that PillarSegNet achieves a
performance gain of about 10% mIoU over the state-of-the-art grid map method.
Related papers
- Bayesian Self-Training for Semi-Supervised 3D Segmentation [59.544558398992386]
3D segmentation is a core problem in computer vision.
densely labeling 3D point clouds to employ fully-supervised training remains too labor intensive and expensive.
Semi-supervised training provides a more practical alternative, where only a small set of labeled data is given, accompanied by a larger unlabeled set.
arXiv Detail & Related papers (2024-09-12T14:54:31Z) - LISNeRF Mapping: LiDAR-based Implicit Mapping via Semantic Neural Fields for Large-Scale 3D Scenes [2.822816116516042]
Large-scale semantic mapping is crucial for outdoor autonomous agents to fulfill high-level tasks such as planning and navigation.
This paper proposes a novel method for large-scale 3D semantic reconstruction through implicit representations from posed LiDAR measurements alone.
arXiv Detail & Related papers (2023-11-04T03:55:38Z) - Volumetric Semantically Consistent 3D Panoptic Mapping [77.13446499924977]
We introduce an online 2D-to-3D semantic instance mapping algorithm aimed at generating semantic 3D maps suitable for autonomous agents in unstructured environments.
It introduces novel ways of integrating semantic prediction confidence during mapping, producing semantic and instance-consistent 3D regions.
The proposed method achieves accuracy superior to the state of the art on public large-scale datasets, improving on a number of widely used metrics.
arXiv Detail & Related papers (2023-09-26T08:03:10Z) - PointOcc: Cylindrical Tri-Perspective View for Point-based 3D Semantic
Occupancy Prediction [72.75478398447396]
We propose a cylindrical tri-perspective view to represent point clouds effectively and comprehensively.
Considering the distance distribution of LiDAR point clouds, we construct the tri-perspective view in the cylindrical coordinate system.
We employ spatial group pooling to maintain structural details during projection and adopt 2D backbones to efficiently process each TPV plane.
arXiv Detail & Related papers (2023-08-31T17:57:17Z) - SATR: Zero-Shot Semantic Segmentation of 3D Shapes [74.08209893396271]
We explore the task of zero-shot semantic segmentation of 3D shapes by using large-scale off-the-shelf 2D image recognition models.
We develop the Assignment with Topological Reweighting (SATR) algorithm and evaluate it on ShapeNetPart and our proposed FAUST benchmarks.
SATR achieves state-of-the-art performance and outperforms a baseline algorithm by 1.3% and 4% average mIoU.
arXiv Detail & Related papers (2023-04-11T00:43:16Z) - Improving Lidar-Based Semantic Segmentation of Top-View Grid Maps by
Learning Features in Complementary Representations [3.0413873719021995]
We introduce a novel way to predict semantic information from sparse, single-shot LiDAR measurements in the context of autonomous driving.
The approach is aimed specifically at improving the semantic segmentation of top-view grid maps.
For each representation a tailored deep learning architecture is developed to effectively extract semantic information.
arXiv Detail & Related papers (2022-03-02T14:49:51Z) - LatticeNet: Fast Spatio-Temporal Point Cloud Segmentation Using
Permutohedral Lattices [27.048998326468688]
Deep convolutional neural networks (CNNs) have shown outstanding performance in the task of semantically segmenting images.
Here, we propose LatticeNet, a novel approach for 3D semantic segmentation, which takes raw point clouds as input.
We present results of 3D segmentation on multiple datasets where our method achieves state-of-the-art performance.
arXiv Detail & Related papers (2021-08-09T10:17:27Z) - S3Net: 3D LiDAR Sparse Semantic Segmentation Network [1.330528227599978]
S3Net is a novel convolutional neural network for LiDAR point cloud semantic segmentation.
It adopts an encoder-decoder backbone that consists of Sparse Intra-channel Attention Module (SIntraAM) and Sparse Inter-channel Attention Module (SInterAM)
arXiv Detail & Related papers (2021-03-15T22:15:24Z) - Semantic Segmentation on Swiss3DCities: A Benchmark Study on Aerial
Photogrammetric 3D Pointcloud Dataset [67.44497676652173]
We introduce a new outdoor urban 3D pointcloud dataset, covering a total area of 2.7 $km2$, sampled from three Swiss cities.
The dataset is manually annotated for semantic segmentation with per-point labels, and is built using photogrammetry from images acquired by multirotors equipped with high-resolution cameras.
arXiv Detail & Related papers (2020-12-23T21:48:47Z) - Radar-based Dynamic Occupancy Grid Mapping and Object Detection [55.74894405714851]
In recent years, the classical occupancy grid map approach has been extended to dynamic occupancy grid maps.
This paper presents the further development of a previous approach.
The data of multiple radar sensors are fused, and a grid-based object tracking and mapping method is applied.
arXiv Detail & Related papers (2020-08-09T09:26:30Z) - Exploiting Multi-Layer Grid Maps for Surround-View Semantic Segmentation
of Sparse LiDAR Data [2.6876976011647145]
We consider the transformation of laser range measurements into a top-view grid map representation to approach the task of LiDAR-only semantic segmentation.
We are exploiting a grid map framework to extract relevant information and represent them by using multi-layer grid maps.
We compare single-layer and multi-layer approaches and demonstrate the benefit of a multi-layer grid map input.
arXiv Detail & Related papers (2020-05-13T23:50:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.