PCSCNet: Fast 3D Semantic Segmentation of LiDAR Point Cloud for
Autonomous Car using Point Convolution and Sparse Convolution Network
- URL: http://arxiv.org/abs/2202.10047v1
- Date: Mon, 21 Feb 2022 08:31:37 GMT
- Title: PCSCNet: Fast 3D Semantic Segmentation of LiDAR Point Cloud for
Autonomous Car using Point Convolution and Sparse Convolution Network
- Authors: Jaehyun Park, Chansoo Kim, Kichun Jo
- Abstract summary: We propose a fast voxel-based semantic segmentation model using Point Convolution and 3D Sparse Convolution (PCSCNet)
The proposed model is designed to outperform at both high and low voxel resolution using point convolution-based feature extraction.
- Score: 8.959391124399925
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The autonomous car must recognize the driving environment quickly for safe
driving. As the Light Detection And Range (LiDAR) sensor is widely used in the
autonomous car, fast semantic segmentation of LiDAR point cloud, which is the
point-wise classification of the point cloud within the sensor framerate, has
attracted attention in recognition of the driving environment. Although the
voxel and fusion-based semantic segmentation models are the state-of-the-art
model in point cloud semantic segmentation recently, their real-time
performance suffer from high computational load due to high voxel resolution.
In this paper, we propose the fast voxel-based semantic segmentation model
using Point Convolution and 3D Sparse Convolution (PCSCNet). The proposed model
is designed to outperform at both high and low voxel resolution using point
convolution-based feature extraction. Moreover, the proposed model accelerates
the feature propagation using 3D sparse convolution after the feature
extraction. The experimental results demonstrate that the proposed model
outperforms the state-of-the-art real-time models in semantic segmentation of
SemanticKITTI and nuScenes, and achieves the real-time performance in LiDAR
point cloud inference.
Related papers
- OPUS: Occupancy Prediction Using a Sparse Set [64.60854562502523]
We present a framework to simultaneously predict occupied locations and classes using a set of learnable queries.
OPUS incorporates a suite of non-trivial strategies to enhance model performance.
Our lightest model achieves superior RayIoU on the Occ3D-nuScenes dataset at near 2x FPS, while our heaviest model surpasses previous best results by 6.1 RayIoU.
arXiv Detail & Related papers (2024-09-14T07:44:22Z) - Future Does Matter: Boosting 3D Object Detection with Temporal Motion Estimation in Point Cloud Sequences [25.74000325019015]
We introduce a novel LiDAR 3D object detection framework, namely LiSTM, to facilitate spatial-temporal feature learning with cross-frame motion forecasting information.
We have conducted experiments on the aggregation and nuScenes datasets to demonstrate that the proposed framework achieves superior 3D detection performance.
arXiv Detail & Related papers (2024-09-06T16:29:04Z) - FASTC: A Fast Attentional Framework for Semantic Traversability Classification Using Point Cloud [7.711666704468952]
We address the problem of traversability assessment using point clouds.
We propose a pillar feature extraction module that utilizes PointNet to capture features from point clouds organized in vertical volume.
We then propose a newtemporal attention module to fuse multi-frame information, which can properly handle the varying density problem of LIDAR point clouds.
arXiv Detail & Related papers (2024-06-24T12:01:55Z) - Rethinking Voxelization and Classification for 3D Object Detection [68.8204255655161]
The main challenge in 3D object detection from LiDAR point clouds is achieving real-time performance without affecting the reliability of the network.
We present a solution to improve network inference speed and precision at the same time by implementing a fast dynamic voxelizer.
In addition, we propose a lightweight detection sub-head model for classifying predicted objects and filter out false detected objects.
arXiv Detail & Related papers (2023-01-10T16:22:04Z) - Analyzing Deep Learning Representations of Point Clouds for Real-Time
In-Vehicle LiDAR Perception [2.365702128814616]
We propose a novel computational taxonomy of LiDAR point cloud representations used in modern deep neural networks for 3D point cloud processing.
Thereby, we uncover common advantages and limitations in terms of computational efficiency, memory requirements, and representational capacity.
arXiv Detail & Related papers (2022-10-26T10:39:59Z) - BEVDetNet: Bird's Eye View LiDAR Point Cloud based Real-time 3D Object
Detection for Autonomous Driving [6.389322215324224]
We propose a novel semantic segmentation architecture as a single unified model for object center detection using key points, box predictions and orientation prediction.
The proposed architecture can be trivially extended to include semantic segmentation classes like road without any additional computation.
The model is 5X faster than other top accuracy models with a minimal accuracy degradation of 2% in Average Precision at IoU=0.5 on KITTI dataset.
arXiv Detail & Related papers (2021-04-21T22:06:39Z) - LiDAR-based Panoptic Segmentation via Dynamic Shifting Network [56.71765153629892]
LiDAR-based panoptic segmentation aims to parse both objects and scenes in a unified manner.
We propose the Dynamic Shifting Network (DS-Net), which serves as an effective panoptic segmentation framework in the point cloud realm.
Our proposed DS-Net achieves superior accuracies over current state-of-the-art methods.
arXiv Detail & Related papers (2020-11-24T08:44:46Z) - Multi Projection Fusion for Real-time Semantic Segmentation of 3D LiDAR
Point Clouds [2.924868086534434]
This paper introduces a novel approach for 3D point cloud semantic segmentation that exploits multiple projections of the point cloud.
Our Multi-Projection Fusion framework analyzes spherical and bird's-eye view projections using two separate highly-efficient 2D fully convolutional models.
arXiv Detail & Related papers (2020-11-03T19:40:43Z) - InfoFocus: 3D Object Detection for Autonomous Driving with Dynamic
Information Modeling [65.47126868838836]
We propose a novel 3D object detection framework with dynamic information modeling.
Coarse predictions are generated in the first stage via a voxel-based region proposal network.
Experiments are conducted on the large-scale nuScenes 3D detection benchmark.
arXiv Detail & Related papers (2020-07-16T18:27:08Z) - Pseudo-LiDAR Point Cloud Interpolation Based on 3D Motion Representation
and Spatial Supervision [68.35777836993212]
We propose a Pseudo-LiDAR point cloud network to generate temporally and spatially high-quality point cloud sequences.
By exploiting the scene flow between point clouds, the proposed network is able to learn a more accurate representation of the 3D spatial motion relationship.
arXiv Detail & Related papers (2020-06-20T03:11:04Z) - Reconfigurable Voxels: A New Representation for LiDAR-Based Point Clouds [76.52448276587707]
We propose Reconfigurable Voxels, a new approach to constructing representations from 3D point clouds.
Specifically, we devise a biased random walk scheme, which adaptively covers each neighborhood with a fixed number of voxels.
We find that this approach effectively improves the stability of voxel features, especially for sparse regions.
arXiv Detail & Related papers (2020-04-06T15:07:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.