MixSup: Mixed-grained Supervision for Label-efficient LiDAR-based 3D
Object Detection
- URL: http://arxiv.org/abs/2401.16305v1
- Date: Mon, 29 Jan 2024 17:05:19 GMT
- Title: MixSup: Mixed-grained Supervision for Label-efficient LiDAR-based 3D
Object Detection
- Authors: Yuxue Yang, Lue Fan, Zhaoxiang Zhang
- Abstract summary: MixSup is a more practical paradigm simultaneously utilizing massive cheap coarse labels and a limited number of accurate labels for Mixed-grained Supervision.
MixSup achieves up to 97.31% of fully supervised performance, using cheap cluster annotations and only 10% box annotations.
- Score: 59.1417156002086
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Label-efficient LiDAR-based 3D object detection is currently dominated by
weakly/semi-supervised methods. Instead of exclusively following one of them,
we propose MixSup, a more practical paradigm simultaneously utilizing massive
cheap coarse labels and a limited number of accurate labels for Mixed-grained
Supervision. We start by observing that point clouds are usually textureless,
making it hard to learn semantics. However, point clouds are geometrically rich
and scale-invariant to the distances from sensors, making it relatively easy to
learn the geometry of objects, such as poses and shapes. Thus, MixSup leverages
massive coarse cluster-level labels to learn semantics and a few expensive
box-level labels to learn accurate poses and shapes. We redesign the label
assignment in mainstream detectors, which allows them seamlessly integrated
into MixSup, enabling practicality and universality. We validate its
effectiveness in nuScenes, Waymo Open Dataset, and KITTI, employing various
detectors. MixSup achieves up to 97.31% of fully supervised performance, using
cheap cluster annotations and only 10% box annotations. Furthermore, we propose
PointSAM based on the Segment Anything Model for automated coarse labeling,
further reducing the annotation burden. The code is available at
https://github.com/BraveGroup/PointSAM-for-MixSup.
Related papers
- Weakly Supervised 3D Instance Segmentation without Instance-level
Annotations [57.615325809883636]
3D semantic scene understanding tasks have achieved great success with the emergence of deep learning, but often require a huge amount of manually annotated training data.
We propose the first weakly-supervised 3D instance segmentation method that only requires categorical semantic labels as supervision.
By generating pseudo instance labels from categorical semantic labels, our designed approach can also assist existing methods for learning 3D instance segmentation at reduced annotation cost.
arXiv Detail & Related papers (2023-08-03T12:30:52Z) - You Only Need One Thing One Click: Self-Training for Weakly Supervised
3D Scene Understanding [107.06117227661204]
We propose One Thing One Click'', meaning that the annotator only needs to label one point per object.
We iteratively conduct the training and label propagation, facilitated by a graph propagation module.
Our model can be compatible to 3D instance segmentation equipped with a point-clustering strategy.
arXiv Detail & Related papers (2023-03-26T13:57:00Z) - MaskCon: Masked Contrastive Learning for Coarse-Labelled Dataset [19.45520684918576]
We propose a contrastive learning method, called $textbfMask$ed $textbfCon$trastive learning($textbfMaskCon$)
For each sample our method generates soft-labels with the aid of coarse labels against other samples and another augmented view of the sample in question.
Our method achieves significant improvement over the current state-of-the-art in various datasets.
arXiv Detail & Related papers (2023-03-22T17:08:31Z) - Box2Seg: Learning Semantics of 3D Point Clouds with Box-Level
Supervision [65.19589997822155]
We introduce a neural architecture, termed Box2Seg, to learn point-level semantics of 3D point clouds with bounding box-level supervision.
We show that the proposed network can be trained with cheap, or even off-the-shelf bounding box-level annotations and subcloud-level tags.
arXiv Detail & Related papers (2022-01-09T09:07:48Z) - Dense Supervision Propagation for Weakly Supervised Semantic Segmentation on 3D Point Clouds [59.63231842439687]
We train a semantic point cloud segmentation network with only a small portion of points being labeled.
We propose a cross-sample feature reallocating module to transfer similar features and therefore re-route the gradients across two samples.
Our weakly supervised method with only 10% and 1% of labels can produce compatible results with the fully supervised counterpart.
arXiv Detail & Related papers (2021-07-23T14:34:57Z) - Multi-Path Region Mining For Weakly Supervised 3D Semantic Segmentation
on Point Clouds [67.0904905172941]
We propose a weakly supervised approach to predict point-level results using weak labels on 3D point clouds.
To the best of our knowledge, this is the first method that uses cloud-level weak labels on raw 3D space to train a point cloud semantic segmentation network.
arXiv Detail & Related papers (2020-03-29T14:13:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.