Few-Shot 3D Point Cloud Semantic Segmentation via Stratified
Class-Specific Attention Based Transformer Network
- URL: http://arxiv.org/abs/2303.15654v1
- Date: Tue, 28 Mar 2023 00:27:54 GMT
- Title: Few-Shot 3D Point Cloud Semantic Segmentation via Stratified
Class-Specific Attention Based Transformer Network
- Authors: Canyu Zhang, Zhenyao Wu, Xinyi Wu, Ziyu Zhao, Song Wang
- Abstract summary: We develop a new multi-layer transformer network for few-shot point cloud semantic segmentation.
Our method achieves the new state-of-the-art performance, with 15% less inference time, over existing few-shot 3D point cloud segmentation models.
- Score: 22.9434434107516
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D point cloud semantic segmentation aims to group all points into different
semantic categories, which benefits important applications such as point cloud
scene reconstruction and understanding. Existing supervised point cloud
semantic segmentation methods usually require large-scale annotated point
clouds for training and cannot handle new categories. While a few-shot learning
method was proposed recently to address these two problems, it suffers from
high computational complexity caused by graph construction and inability to
learn fine-grained relationships among points due to the use of pooling
operations. In this paper, we further address these problems by developing a
new multi-layer transformer network for few-shot point cloud semantic
segmentation. In the proposed network, the query point cloud features are
aggregated based on the class-specific support features in different scales.
Without using pooling operations, our method makes full use of all pixel-level
features from the support samples. By better leveraging the support features
for few-shot learning, the proposed method achieves the new state-of-the-art
performance, with 15\% less inference time, over existing few-shot 3D point
cloud segmentation models on the S3DIS dataset and the ScanNet dataset.
Related papers
- Generalized Few-Shot Point Cloud Segmentation Via Geometric Words [54.32239996417363]
Few-shot point cloud segmentation algorithms learn to adapt to new classes at the sacrifice of segmentation accuracy for the base classes.
We present the first attempt at a more practical paradigm of generalized few-shot point cloud segmentation.
We propose the geometric words to represent geometric components shared between the base and novel classes, and incorporate them into a novel geometric-aware semantic representation.
arXiv Detail & Related papers (2023-09-20T11:24:33Z) - Clustering based Point Cloud Representation Learning for 3D Analysis [80.88995099442374]
We propose a clustering based supervised learning scheme for point cloud analysis.
Unlike current de-facto, scene-wise training paradigm, our algorithm conducts within-class clustering on the point embedding space.
Our algorithm shows notable improvements on famous point cloud segmentation datasets.
arXiv Detail & Related papers (2023-07-27T03:42:12Z) - Prototype Adaption and Projection for Few- and Zero-shot 3D Point Cloud
Semantic Segmentation [30.18333233940194]
We address the challenging task of few-shot and zero-shot 3D point cloud semantic segmentation.
Our proposed method surpasses state-of-the-art algorithms by a considerable 7.90% and 14.82% under the 2-way 1-shot setting on S3DIS and ScanNet benchmarks, respectively.
arXiv Detail & Related papers (2023-05-23T17:58:05Z) - FreePoint: Unsupervised Point Cloud Instance Segmentation [72.64540130803687]
We propose FreePoint, for underexplored unsupervised class-agnostic instance segmentation on point clouds.
We represent point features by combining coordinates, colors, and self-supervised deep features.
Based on the point features, we segment point clouds into coarse instance masks as pseudo labels, which are used to train a point cloud instance segmentation model.
arXiv Detail & Related papers (2023-05-11T16:56:26Z) - DRINet: A Dual-Representation Iterative Learning Network for Point Cloud
Segmentation [45.768040873409824]
DRINet serves as the basic network structure for dual-representation learning.
Our network achieves state-of-the-art results for point cloud classification and segmentation tasks.
For large-scale outdoor scenarios, our method outperforms state-of-the-art methods with a real-time inference speed of 62ms per frame.
arXiv Detail & Related papers (2021-08-09T13:23:54Z) - Compositional Prototype Network with Multi-view Comparision for Few-Shot
Point Cloud Semantic Segmentation [47.0611707526858]
A fully supervised point cloud segmentation network often requires a large amount of data with point-wise annotations.
We present the Compositional Prototype Network that can undertake point cloud segmentation with only a few labeled training data.
Inspired by the few-shot learning literature in images, our network directly transfers label information from the limited training data to unlabeled test data for prediction.
arXiv Detail & Related papers (2020-12-28T15:01:34Z) - Few-shot 3D Point Cloud Semantic Segmentation [138.80825169240302]
We propose a novel attention-aware multi-prototype transductive few-shot point cloud semantic segmentation method.
Our proposed method shows significant and consistent improvements compared to baselines in different few-shot point cloud semantic segmentation settings.
arXiv Detail & Related papers (2020-06-22T08:05:25Z) - GRNet: Gridding Residual Network for Dense Point Cloud Completion [54.43648460932248]
Estimating the complete 3D point cloud from an incomplete one is a key problem in many vision and robotics applications.
We propose a novel Gridding Residual Network (GRNet) for point cloud completion.
Experimental results indicate that the proposed GRNet performs favorably against state-of-the-art methods on the ShapeNet, Completion3D, and KITTI benchmarks.
arXiv Detail & Related papers (2020-06-06T02:46:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.