What Makes for Effective Few-shot Point Cloud Classification?
- URL: http://arxiv.org/abs/2304.00022v1
- Date: Fri, 31 Mar 2023 15:55:06 GMT
- Title: What Makes for Effective Few-shot Point Cloud Classification?
- Authors: Chuangguan Ye, Hongyuan Zhu, Yongbin Liao, Yanggang Zhang, Tao Chen,
Jiayuan Fan
- Abstract summary: We show that 3D few-shot learning is more challenging with unordered structures, high intra-class variances, and subtle inter-class differences.
We propose a novel plug-and-play component called Cross-Instance Adaptation (CIA) module, to address the high intra-class variances and subtle inter-class differences issues.
- Score: 18.62689395276194
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to the emergence of powerful computing resources and large-scale
annotated datasets, deep learning has seen wide applications in our daily life.
However, most current methods require extensive data collection and retraining
when dealing with novel classes never seen before. On the other hand, we humans
can quickly recognize new classes by looking at a few samples, which motivates
the recent popularity of few-shot learning (FSL) in machine learning
communities. Most current FSL approaches work on 2D image domain, however, its
implication in 3D perception is relatively under-explored. Not only needs to
recognize the unseen examples as in 2D domain, 3D few-shot learning is more
challenging with unordered structures, high intra-class variances, and subtle
inter-class differences. Moreover, different architectures and learning
algorithms make it difficult to study the effectiveness of existing 2D methods
when migrating to the 3D domain. In this work, for the first time, we perform
systematic and extensive studies of recent 2D FSL and 3D backbone networks for
benchmarking few-shot point cloud classification, and we suggest a strong
baseline and learning architectures for 3D FSL. Then, we propose a novel
plug-and-play component called Cross-Instance Adaptation (CIA) module, to
address the high intra-class variances and subtle inter-class differences
issues, which can be easily inserted into current baselines with significant
performance improvement. Extensive experiments on two newly introduced
benchmark datasets, ModelNet40-FS and ShapeNet70-FS, demonstrate the
superiority of our proposed network for 3D FSL.
Related papers
- Foundation Model-Powered 3D Few-Shot Class Incremental Learning via Training-free Adaptor [9.54964908165465]
This paper introduces a new method to tackle the Few-Shot Continual Incremental Learning problem in 3D point cloud environments.
We leverage a foundational 3D model trained extensively on point cloud data.
Our approach uses a dual cache system: first, it uses previous test samples based on how confident the model was in its predictions to prevent forgetting, and second, it includes a small number of new task samples to prevent overfitting.
arXiv Detail & Related papers (2024-10-11T20:23:00Z) - OV-Uni3DETR: Towards Unified Open-Vocabulary 3D Object Detection via Cycle-Modality Propagation [67.56268991234371]
OV-Uni3DETR achieves the state-of-the-art performance on various scenarios, surpassing existing methods by more than 6% on average.
Code and pre-trained models will be released later.
arXiv Detail & Related papers (2024-03-28T17:05:04Z) - Leveraging Large-Scale Pretrained Vision Foundation Models for
Label-Efficient 3D Point Cloud Segmentation [67.07112533415116]
We present a novel framework that adapts various foundational models for the 3D point cloud segmentation task.
Our approach involves making initial predictions of 2D semantic masks using different large vision models.
To generate robust 3D semantic pseudo labels, we introduce a semantic label fusion strategy that effectively combines all the results via voting.
arXiv Detail & Related papers (2023-11-03T15:41:15Z) - A Closer Look at Few-Shot 3D Point Cloud Classification [21.57893885371941]
We propose a new network, Point-cloud Correlation Interaction ( PCIA), with three novel plug-and-play components called Salient-Part Fusion (SPF), Self-Channel Interaction Plus (SCI+) module, and Cross-Instance Fusion Plus (CIF+) module.
These modules can be inserted into most FSL algorithms with minor changes and significantly improve the performance.
Experimental results on three benchmark datasets, ModelNet40-FS, ShapeNet70-FS, and ScanObjectNN-FS, demonstrate that our method achieves state-of-the-art performance.
arXiv Detail & Related papers (2023-03-31T17:01:13Z) - ULIP: Learning a Unified Representation of Language, Images, and Point
Clouds for 3D Understanding [110.07170245531464]
Current 3D models are limited by datasets with a small number of annotated data and a pre-defined set of categories.
Recent advances have shown that similar problems can be significantly alleviated by employing knowledge from other modalities, such as language.
We learn a unified representation of images, texts, and 3D point clouds by pre-training with object triplets from the three modalities.
arXiv Detail & Related papers (2022-12-10T01:34:47Z) - SSMTL++: Revisiting Self-Supervised Multi-Task Learning for Video
Anomaly Detection [108.57862846523858]
We revisit the self-supervised multi-task learning framework, proposing several updates to the original method.
We modernize the 3D convolutional backbone by introducing multi-head self-attention modules.
In our attempt to further improve the model, we study additional self-supervised learning tasks, such as predicting segmentation maps.
arXiv Detail & Related papers (2022-07-16T19:25:41Z) - Multimodal Semi-Supervised Learning for 3D Objects [19.409295848915388]
This paper explores how the coherence of different modelities of 3D data can be used to improve data efficiency for both 3D classification and retrieval tasks.
We propose a novel multimodal semi-supervised learning framework by introducing instance-level consistency constraint and a novel multimodal contrastive prototype (M2CP) loss.
Our proposed framework significantly outperforms all the state-of-the-art counterparts for both classification and retrieval tasks by a large margin on the modelNet10 and ModelNet40 datasets.
arXiv Detail & Related papers (2021-10-22T05:33:16Z) - Point Discriminative Learning for Unsupervised Representation Learning
on 3D Point Clouds [54.31515001741987]
We propose a point discriminative learning method for unsupervised representation learning on 3D point clouds.
We achieve this by imposing a novel point discrimination loss on the middle level and global level point features.
Our method learns powerful representations and achieves new state-of-the-art performance.
arXiv Detail & Related papers (2021-08-04T15:11:48Z) - Learning without Forgetting for 3D Point Cloud Objects [7.761414660999872]
When we fine-tune a well-trained deep learning model for a new set of classes, the network learns new concepts but gradually forgets the knowledge of old training.
We investigate knowledge distillation techniques on 3D data to reduce catastrophic forgetting of the previous training.
We observe that exploring the interrelation of old and new knowledge during training helps to learn new concepts without forgetting old ones.
arXiv Detail & Related papers (2021-06-27T16:39:39Z) - Zero-Shot Learning on 3D Point Cloud Objects and Beyond [21.6491982908705]
We identify some of the challenges and apply 2D Zero-Shot Learning (ZSL) methods in the 3D domain to analyze the performance of existing models.
A novel loss function is developed that simultaneously aligns seen semantics with point cloud features.
An extensive set of experiments is carried out, establishing state-of-the-art for ZSL and GZSL on synthetic and real datasets.
arXiv Detail & Related papers (2021-04-11T10:04:06Z) - PointContrast: Unsupervised Pre-training for 3D Point Cloud
Understanding [107.02479689909164]
In this work, we aim at facilitating research on 3D representation learning.
We measure the effect of unsupervised pre-training on a large source set of 3D scenes.
arXiv Detail & Related papers (2020-07-21T17:59:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.