Image Understands Point Cloud: Weakly Supervised 3D Semantic
Segmentation via Association Learning
- URL: http://arxiv.org/abs/2209.07774v1
- Date: Fri, 16 Sep 2022 07:59:04 GMT
- Title: Image Understands Point Cloud: Weakly Supervised 3D Semantic
Segmentation via Association Learning
- Authors: Tianfang Sun, Zhizhong Zhang, Xin Tan, Yanyun Qu, Yuan Xie, Lizhuang
Ma
- Abstract summary: We propose a novel cross-modality weakly supervised method for 3D segmentation, incorporating complementary information from unlabeled images.
Basically, we design a dual-branch network equipped with an active labeling strategy, to maximize the power of tiny parts of labels.
Our method even outperforms the state-of-the-art fully supervised competitors with less than 1% actively selected annotations.
- Score: 59.64695628433855
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Weakly supervised point cloud semantic segmentation methods that require 1\%
or fewer labels, hoping to realize almost the same performance as fully
supervised approaches, which recently, have attracted extensive research
attention. A typical solution in this framework is to use self-training or
pseudo labeling to mine the supervision from the point cloud itself, but ignore
the critical information from images. In fact, cameras widely exist in LiDAR
scenarios and this complementary information seems to be greatly important for
3D applications. In this paper, we propose a novel cross-modality weakly
supervised method for 3D segmentation, incorporating complementary information
from unlabeled images. Basically, we design a dual-branch network equipped with
an active labeling strategy, to maximize the power of tiny parts of labels and
directly realize 2D-to-3D knowledge transfer. Afterwards, we establish a
cross-modal self-training framework in an Expectation-Maximum (EM) perspective,
which iterates between pseudo labels estimation and parameters updating. In the
M-Step, we propose a cross-modal association learning to mine complementary
supervision from images by reinforcing the cycle-consistency between 3D points
and 2D superpixels. In the E-step, a pseudo label self-rectification mechanism
is derived to filter noise labels thus providing more accurate labels for the
networks to get fully trained. The extensive experimental results demonstrate
that our method even outperforms the state-of-the-art fully supervised
competitors with less than 1\% actively selected annotations.
Related papers
- Beyond the Label Itself: Latent Labels Enhance Semi-supervised Point
Cloud Panoptic Segmentation [46.01433705072047]
We find two types of latent labels behind the displayed label embedded in LiDAR and image data.
We propose a novel augmentation, Cylinder-Mix, which is able to augment more yet reliable samples for training.
We also propose the Instance Position-scale Learning (IPSL) Module to learn and fuse the information of instance position and scale.
arXiv Detail & Related papers (2023-12-13T15:56:24Z) - Cross-modal and Cross-domain Knowledge Transfer for Label-free 3D
Segmentation [23.110443633049382]
We propose a novel approach for the challenging cross-modal and cross-domain adaptation task by fully exploring the relationship between images and point clouds.
Our method achieves state-of-the-art performance for 3D point cloud semantic segmentation on Semantic KITTI by using the knowledge of KITTI360 and GTA5.
arXiv Detail & Related papers (2023-09-19T14:29:57Z) - Towards Label-free Scene Understanding by Vision Foundation Models [87.13117617056004]
We investigate the potential of vision foundation models in enabling networks to comprehend 2D and 3D worlds without labelled data.
We propose a novel Cross-modality Noisy Supervision (CNS) method that leverages the strengths of CLIP and SAM to supervise 2D and 3D networks simultaneously.
Our 2D and 3D network achieves label-free semantic segmentation with 28.4% and 33.5% mIoU on ScanNet, improving 4.7% and 7.9%, respectively.
arXiv Detail & Related papers (2023-06-06T17:57:49Z) - You Only Need One Thing One Click: Self-Training for Weakly Supervised
3D Scene Understanding [107.06117227661204]
We propose One Thing One Click'', meaning that the annotator only needs to label one point per object.
We iteratively conduct the training and label propagation, facilitated by a graph propagation module.
Our model can be compatible to 3D instance segmentation equipped with a point-clustering strategy.
arXiv Detail & Related papers (2023-03-26T13:57:00Z) - Collaborative Propagation on Multiple Instance Graphs for 3D Instance
Segmentation with Single-point Supervision [63.429704654271475]
We propose a novel weakly supervised method RWSeg that only requires labeling one object with one point.
With these sparse weak labels, we introduce a unified framework with two branches to propagate semantic and instance information.
Specifically, we propose a Cross-graph Competing Random Walks (CRW) algorithm that encourages competition among different instance graphs.
arXiv Detail & Related papers (2022-08-10T02:14:39Z) - Dense Supervision Propagation for Weakly Supervised Semantic Segmentation on 3D Point Clouds [59.63231842439687]
We train a semantic point cloud segmentation network with only a small portion of points being labeled.
We propose a cross-sample feature reallocating module to transfer similar features and therefore re-route the gradients across two samples.
Our weakly supervised method with only 10% and 1% of labels can produce compatible results with the fully supervised counterpart.
arXiv Detail & Related papers (2021-07-23T14:34:57Z) - One Thing One Click: A Self-Training Approach for Weakly Supervised 3D
Semantic Segmentation [78.36781565047656]
We propose "One Thing One Click," meaning that the annotator only needs to label one point per object.
We iteratively conduct the training and label propagation, facilitated by a graph propagation module.
Our results are also comparable to those of the fully supervised counterparts.
arXiv Detail & Related papers (2021-04-06T02:27:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.