Omni-supervised Point Cloud Segmentation via Gradual Receptive Field
Component Reasoning
- URL: http://arxiv.org/abs/2105.10203v1
- Date: Fri, 21 May 2021 08:32:02 GMT
- Title: Omni-supervised Point Cloud Segmentation via Gradual Receptive Field
Component Reasoning
- Authors: Jingyu Gong, Jiachen Xu, Xin Tan, Haichuan Song, Yanyun Qu, Yuan Xie,
Lizhuang Ma
- Abstract summary: We bring the first omni-scale supervision method to point cloud segmentation via the proposed gradual Receptive Field Component Reasoning (RFCR)
Our method brings new state-of-the-art performances for S3DIS as well as Semantic3D and ranks the 1st in the ScanNet benchmark among all the point-based methods.
- Score: 41.83979510282989
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hidden features in neural network usually fail to learn informative
representation for 3D segmentation as supervisions are only given on output
prediction, while this can be solved by omni-scale supervision on intermediate
layers. In this paper, we bring the first omni-scale supervision method to
point cloud segmentation via the proposed gradual Receptive Field Component
Reasoning (RFCR), where target Receptive Field Component Codes (RFCCs) are
designed to record categories within receptive fields for hidden units in the
encoder. Then, target RFCCs will supervise the decoder to gradually infer the
RFCCs in a coarse-to-fine categories reasoning manner, and finally obtain the
semantic labels. Because many hidden features are inactive with tiny magnitude
and make minor contributions to RFCC prediction, we propose a Feature
Densification with a centrifugal potential to obtain more unambiguous features,
and it is in effect equivalent to entropy regularization over features. More
active features can further unleash the potential of our omni-supervision
method. We embed our method into four prevailing backbones and test on three
challenging benchmarks. Our method can significantly improve the backbones in
all three datasets. Specifically, our method brings new state-of-the-art
performances for S3DIS as well as Semantic3D and ranks the 1st in the ScanNet
benchmark among all the point-based methods. Code will be publicly available at
https://github.com/azuki-miho/RFCR.
Related papers
- Subspace Prototype Guidance for Mitigating Class Imbalance in Point Cloud Semantic Segmentation [23.250178208474928]
This paper introduces a novel method, namely subspace prototype guidance (textbfSPG) to guide the training of segmentation network.
The proposed method significantly improves the segmentation performance and surpasses the state-of-the-art method.
arXiv Detail & Related papers (2024-08-20T04:31:46Z) - FASTC: A Fast Attentional Framework for Semantic Traversability Classification Using Point Cloud [7.711666704468952]
We address the problem of traversability assessment using point clouds.
We propose a pillar feature extraction module that utilizes PointNet to capture features from point clouds organized in vertical volume.
We then propose a newtemporal attention module to fuse multi-frame information, which can properly handle the varying density problem of LIDAR point clouds.
arXiv Detail & Related papers (2024-06-24T12:01:55Z) - Multi-modality Affinity Inference for Weakly Supervised 3D Semantic
Segmentation [47.81638388980828]
We propose a simple yet effective scene-level weakly supervised point cloud segmentation method with a newly introduced multi-modality point affinity inference module.
Our method outperforms the state-of-the-art by 4% to 6% mIoU on the ScanNet and S3DIS benchmarks.
arXiv Detail & Related papers (2023-12-27T14:01:35Z) - Learning Latent Part-Whole Hierarchies for Point Clouds [41.288934432615676]
We propose an encoder-decoder style latent variable model that explicitly learns the part-whole hierarchies for the point cloud segmentation.
The proposed method achieves state-of-the-art performance in not only top-level part segmentation but also middle-level latent subpart segmentation.
arXiv Detail & Related papers (2022-11-14T03:17:33Z) - CloudAttention: Efficient Multi-Scale Attention Scheme For 3D Point
Cloud Learning [81.85951026033787]
We set transformers in this work and incorporate them into a hierarchical framework for shape classification and part and scene segmentation.
We also compute efficient and dynamic global cross attentions by leveraging sampling and grouping at each iteration.
The proposed hierarchical model achieves state-of-the-art shape classification in mean accuracy and yields results on par with the previous segmentation methods.
arXiv Detail & Related papers (2022-07-31T21:39:15Z) - Dense Supervision Propagation for Weakly Supervised Semantic Segmentation on 3D Point Clouds [59.63231842439687]
We train a semantic point cloud segmentation network with only a small portion of points being labeled.
We propose a cross-sample feature reallocating module to transfer similar features and therefore re-route the gradients across two samples.
Our weakly supervised method with only 10% and 1% of labels can produce compatible results with the fully supervised counterpart.
arXiv Detail & Related papers (2021-07-23T14:34:57Z) - (AF)2-S3Net: Attentive Feature Fusion with Adaptive Feature Selection
for Sparse Semantic Segmentation Network [3.6967381030744515]
We propose AF2-S3Net, an end-to-end encoder-decoder CNN network for 3D LiDAR semantic segmentation.
We present a novel multi-branch attentive feature fusion module in the encoder and a unique adaptive feature selection module with feature map re-weighting in the decoder.
Our experimental results show that the proposed method outperforms the state-of-the-art approaches on the large-scale SemanticKITTI benchmark.
arXiv Detail & Related papers (2021-02-08T21:04:21Z) - End-to-End Object Detection with Fully Convolutional Network [71.56728221604158]
We introduce a Prediction-aware One-To-One (POTO) label assignment for classification to enable end-to-end detection.
A simple 3D Max Filtering (3DMF) is proposed to utilize the multi-scale features and improve the discriminability of convolutions in the local region.
Our end-to-end framework achieves competitive performance against many state-of-the-art detectors with NMS on COCO and CrowdHuman datasets.
arXiv Detail & Related papers (2020-12-07T09:14:55Z) - Unsupervised Feedforward Feature (UFF) Learning for Point Cloud
Classification and Segmentation [57.62713515497585]
Unsupervised feedforward feature learning is proposed for joint classification and segmentation of 3D point clouds.
The UFF method exploits statistical correlations of points in a point cloud set to learn shape and point features in a one-pass feedforward manner.
It learns global shape features through the encoder and local point features through the encoder-decoder architecture.
arXiv Detail & Related papers (2020-09-02T18:25:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.