Beyond the Label Itself: Latent Labels Enhance Semi-supervised Point
Cloud Panoptic Segmentation
- URL: http://arxiv.org/abs/2312.08234v2
- Date: Sun, 11 Feb 2024 12:19:08 GMT
- Title: Beyond the Label Itself: Latent Labels Enhance Semi-supervised Point
Cloud Panoptic Segmentation
- Authors: Yujun Chen, Xin Tan, Zhizhong Zhang, Yanyun Qu, Yuan Xie
- Abstract summary: We find two types of latent labels behind the displayed label embedded in LiDAR and image data.
We propose a novel augmentation, Cylinder-Mix, which is able to augment more yet reliable samples for training.
We also propose the Instance Position-scale Learning (IPSL) Module to learn and fuse the information of instance position and scale.
- Score: 46.01433705072047
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As the exorbitant expense of labeling autopilot datasets and the growing
trend of utilizing unlabeled data, semi-supervised segmentation on point clouds
becomes increasingly imperative. Intuitively, finding out more ``unspoken
words'' (i.e., latent instance information) beyond the label itself should be
helpful to improve performance. In this paper, we discover two types of latent
labels behind the displayed label embedded in LiDAR and image data. First, in
the LiDAR Branch, we propose a novel augmentation, Cylinder-Mix, which is able
to augment more yet reliable samples for training. Second, in the Image Branch,
we propose the Instance Position-scale Learning (IPSL) Module to learn and fuse
the information of instance position and scale, which is from a 2D pre-trained
detector and a type of latent label obtained from 3D to 2D projection. Finally,
the two latent labels are embedded into the multi-modal panoptic segmentation
network. The ablation of the IPSL module demonstrates its robust adaptability,
and the experiments evaluated on SemanticKITTI and nuScenes demonstrate that
our model outperforms the state-of-the-art method, LaserMix.
Related papers
- You Only Need One Thing One Click: Self-Training for Weakly Supervised
3D Scene Understanding [107.06117227661204]
We propose One Thing One Click'', meaning that the annotator only needs to label one point per object.
We iteratively conduct the training and label propagation, facilitated by a graph propagation module.
Our model can be compatible to 3D instance segmentation equipped with a point-clustering strategy.
arXiv Detail & Related papers (2023-03-26T13:57:00Z) - LWSIS: LiDAR-guided Weakly Supervised Instance Segmentation for
Autonomous Driving [34.119642131912485]
We present a more artful framework, LiDAR-guided Weakly Supervised Instance (LWSIS)
LWSIS uses the off-the-shelf 3D data, i.e., Point Cloud, together with the 3D boxes, as natural weak supervisions for training the 2D image instance segmentation models.
Our LWSIS not only exploits the complementary information in multimodal data during training, but also significantly reduces the cost of the dense 2D masks.
arXiv Detail & Related papers (2022-12-07T08:08:01Z) - LESS: Label-Efficient Semantic Segmentation for LiDAR Point Clouds [62.49198183539889]
We propose a label-efficient semantic segmentation pipeline for outdoor scenes with LiDAR point clouds.
Our method co-designs an efficient labeling process with semi/weakly supervised learning.
Our proposed method is even highly competitive compared to the fully supervised counterpart with 100% labels.
arXiv Detail & Related papers (2022-10-14T19:13:36Z) - Image Understands Point Cloud: Weakly Supervised 3D Semantic
Segmentation via Association Learning [59.64695628433855]
We propose a novel cross-modality weakly supervised method for 3D segmentation, incorporating complementary information from unlabeled images.
Basically, we design a dual-branch network equipped with an active labeling strategy, to maximize the power of tiny parts of labels.
Our method even outperforms the state-of-the-art fully supervised competitors with less than 1% actively selected annotations.
arXiv Detail & Related papers (2022-09-16T07:59:04Z) - Dense Supervision Propagation for Weakly Supervised Semantic Segmentation on 3D Point Clouds [59.63231842439687]
We train a semantic point cloud segmentation network with only a small portion of points being labeled.
We propose a cross-sample feature reallocating module to transfer similar features and therefore re-route the gradients across two samples.
Our weakly supervised method with only 10% and 1% of labels can produce compatible results with the fully supervised counterpart.
arXiv Detail & Related papers (2021-07-23T14:34:57Z) - One Thing One Click: A Self-Training Approach for Weakly Supervised 3D
Semantic Segmentation [78.36781565047656]
We propose "One Thing One Click," meaning that the annotator only needs to label one point per object.
We iteratively conduct the training and label propagation, facilitated by a graph propagation module.
Our results are also comparable to those of the fully supervised counterparts.
arXiv Detail & Related papers (2021-04-06T02:27:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.