LaserMix for Semi-Supervised LiDAR Semantic Segmentation
- URL: http://arxiv.org/abs/2207.00026v4
- Date: Fri, 1 Sep 2023 09:58:15 GMT
- Title: LaserMix for Semi-Supervised LiDAR Semantic Segmentation
- Authors: Lingdong Kong and Jiawei Ren and Liang Pan and Ziwei Liu
- Abstract summary: We study the underexplored semi-supervised learning (SSL) in LiDAR segmentation.
Our core idea is to leverage the strong spatial cues of LiDAR point clouds to better exploit unlabeled data.
We propose LaserMix to mix laser beams from different LiDAR scans, and then encourage the model to make consistent and confident predictions.
- Score: 56.73779694312137
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Densely annotating LiDAR point clouds is costly, which restrains the
scalability of fully-supervised learning methods. In this work, we study the
underexplored semi-supervised learning (SSL) in LiDAR segmentation. Our core
idea is to leverage the strong spatial cues of LiDAR point clouds to better
exploit unlabeled data. We propose LaserMix to mix laser beams from different
LiDAR scans, and then encourage the model to make consistent and confident
predictions before and after mixing. Our framework has three appealing
properties: 1) Generic: LaserMix is agnostic to LiDAR representations (e.g.,
range view and voxel), and hence our SSL framework can be universally applied.
2) Statistically grounded: We provide a detailed analysis to theoretically
explain the applicability of the proposed framework. 3) Effective:
Comprehensive experimental analysis on popular LiDAR segmentation datasets
(nuScenes, SemanticKITTI, and ScribbleKITTI) demonstrates our effectiveness and
superiority. Notably, we achieve competitive results over fully-supervised
counterparts with 2x to 5x fewer labels and improve the supervised-only
baseline significantly by 10.8% on average. We hope this concise yet
high-performing framework could facilitate future research in semi-supervised
LiDAR segmentation. Code is publicly available.
Related papers
- Multi-Modal Data-Efficient 3D Scene Understanding for Autonomous Driving [58.16024314532443]
We introduce LaserMix++, a framework that integrates laser beam manipulations from disparate LiDAR scans and incorporates LiDAR-camera correspondences to assist data-efficient learning.
Results demonstrate that LaserMix++ outperforms fully supervised alternatives, achieving comparable accuracy with five times fewer annotations.
This substantial advancement underscores the potential of semi-supervised approaches in reducing the reliance on extensive labeled data in LiDAR-based 3D scene understanding systems.
arXiv Detail & Related papers (2024-05-08T17:59:53Z) - Just Add $100 More: Augmenting NeRF-based Pseudo-LiDAR Point Cloud for Resolving Class-imbalance Problem [12.26293873825084]
We propose to leverage pseudo-LiDAR point clouds generated from videos capturing a surround view of miniatures or real-world objects of minor classes.
Our method, called Pseudo Ground Truth Augmentation (PGT-Aug), consists of three main steps: (i) volumetric 3D instance reconstruction using a 2D-to-3D view synthesis model, (ii) object-level domain alignment with LiDAR intensity estimation, and (iii) a hybrid context-aware placement method from ground and map information.
arXiv Detail & Related papers (2024-03-18T08:50:04Z) - UltraLiDAR: Learning Compact Representations for LiDAR Completion and
Generation [51.443788294845845]
We present UltraLiDAR, a data-driven framework for scene-level LiDAR completion, LiDAR generation, and LiDAR manipulation.
We show that by aligning the representation of a sparse point cloud to that of a dense point cloud, we can densify the sparse point clouds.
By learning a prior over the discrete codebook, we can generate diverse, realistic LiDAR point clouds for self-driving.
arXiv Detail & Related papers (2023-11-02T17:57:03Z) - Point-to-Voxel Knowledge Distillation for LiDAR Semantic Segmentation [74.67594286008317]
This article addresses the problem of distilling knowledge from a large teacher model to a slim student network for LiDAR semantic segmentation.
We propose the Point-to-Voxel Knowledge Distillation (PVD), which transfers the hidden knowledge from both point level and voxel level.
arXiv Detail & Related papers (2022-06-05T05:28:32Z) - LiDAR Distillation: Bridging the Beam-Induced Domain Gap for 3D Object
Detection [96.63947479020631]
In many real-world applications, the LiDAR points used by mass-produced robots and vehicles usually have fewer beams than that in large-scale public datasets.
We propose the LiDAR Distillation to bridge the domain gap induced by different LiDAR beams for 3D object detection.
arXiv Detail & Related papers (2022-03-28T17:59:02Z) - End-To-End Optimization of LiDAR Beam Configuration for 3D Object
Detection and Localization [87.56144220508587]
We take a new route to learn to optimize the LiDAR beam configuration for a given application.
We propose a reinforcement learning-based learning-to-optimize framework to automatically optimize the beam configuration.
Our method is especially useful when a low-resolution (low-cost) LiDAR is needed.
arXiv Detail & Related papers (2022-01-11T09:46:31Z) - Scan-based Semantic Segmentation of LiDAR Point Clouds: An Experimental
Study [2.6205925938720833]
State of the art methods use deep neural networks to predict semantic classes for each point in a LiDAR scan.
A powerful and efficient way to process LiDAR measurements is to use two-dimensional, image-like projections.
We demonstrate various techniques to boost the performance and to improve runtime as well as memory constraints.
arXiv Detail & Related papers (2020-04-06T11:08:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.