Explore In-Context Learning for 3D Point Cloud Understanding
- URL: http://arxiv.org/abs/2306.08659v2
- Date: Wed, 27 Dec 2023 09:02:20 GMT
- Title: Explore In-Context Learning for 3D Point Cloud Understanding
- Authors: Zhongbin Fang, Xiangtai Li, Xia Li, Joachim M. Buhmann, Chen Change
Loy, Mengyuan Liu
- Abstract summary: We introduce a novel framework, named Point-In-Context, designed especially for in-context learning in 3D point clouds.
We propose the Joint Sampling module, carefully designed to work in tandem with the general point sampling operator.
We conduct extensive experiments to validate the versatility and adaptability of our proposed methods in handling a wide range of tasks.
- Score: 71.20912026561484
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the rise of large-scale models trained on broad data, in-context
learning has become a new learning paradigm that has demonstrated significant
potential in natural language processing and computer vision tasks. Meanwhile,
in-context learning is still largely unexplored in the 3D point cloud domain.
Although masked modeling has been successfully applied for in-context learning
in 2D vision, directly extending it to 3D point clouds remains a formidable
challenge. In the case of point clouds, the tokens themselves are the point
cloud positions (coordinates) that are masked during inference. Moreover,
position embedding in previous works may inadvertently introduce information
leakage. To address these challenges, we introduce a novel framework, named
Point-In-Context, designed especially for in-context learning in 3D point
clouds, where both inputs and outputs are modeled as coordinates for each task.
Additionally, we propose the Joint Sampling module, carefully designed to work
in tandem with the general point sampling operator, effectively resolving the
aforementioned technical issues. We conduct extensive experiments to validate
the versatility and adaptability of our proposed methods in handling a wide
range of tasks.
Related papers
- A comprehensive overview of deep learning techniques for 3D point cloud classification and semantic segmentation [0.20649496811699863]
This paper analyzes recent progress in deep learning methods employed for point cloud processing.
It serves as a comprehensive review on two major tasks in 3D point cloud processing-- namely, 3D shape classification and semantic segmentation.
arXiv Detail & Related papers (2024-05-20T09:33:27Z) - Point-In-Context: Understanding Point Cloud via In-Context Learning [67.20277182808992]
We introduce Point-In-Context (PIC), a novel framework for 3D point cloud understanding via in-context learning.
We address the technical challenge of effectively extending masked point modeling to 3D point clouds by introducing a Joint Sampling module.
We propose two novel training strategies, In-Context Labeling and In-Context Enhancing, forming an extended version of PIC named Point-In-Context-Segmenter (PIC-S)
arXiv Detail & Related papers (2024-04-18T17:32:32Z) - Language-Assisted 3D Scene Understanding [17.663583203177197]
We propose a language-assisted approach to point cloud feature learning (LAST-PCL)
We achieve de-redundancy and feature dimensionality reduction without compromising textual priors.
The proposed method learns semantically meaningful point cloud features and achieves state-of-the-art or comparable performance in 3D semantic segmentation, 3D object detection, and 3D scene classification tasks.
arXiv Detail & Related papers (2023-12-18T18:54:56Z) - Self-supervised Learning for Pre-Training 3D Point Clouds: A Survey [25.51613543480276]
Self-supervised point cloud representation learning has attracted increasing attention in recent years.
This paper presents a comprehensive survey of self-supervised point cloud representation learning using DNNs.
arXiv Detail & Related papers (2023-05-08T13:20:55Z) - Point2Vec for Self-Supervised Representation Learning on Point Clouds [66.53955515020053]
We extend data2vec to the point cloud domain and report encouraging results on several downstream tasks.
We propose point2vec, which unleashes the full potential of data2vec-like pre-training on point clouds.
arXiv Detail & Related papers (2023-03-29T10:08:29Z) - PointVST: Self-Supervised Pre-training for 3D Point Clouds via
View-Specific Point-to-Image Translation [64.858505571083]
This paper proposes a translative pre-training framework, namely PointVST.
It is driven by a novel self-supervised pretext task of cross-modal translation from 3D point clouds to their corresponding diverse forms of 2D rendered images.
arXiv Detail & Related papers (2022-12-29T07:03:29Z) - Self-Ensemling for 3D Point Cloud Domain Adaption [29.330315360307374]
We propose an end-to-end self-ensembling network (SEN) for 3D point cloud domain adaption tasks.
Our SEN resorts to the advantages of Mean Teacher and semi-supervised learning, and introduces a soft classification loss and a consistency loss.
Our SEN outperforms the state-of-the-art methods on both classification and segmentation tasks.
arXiv Detail & Related papers (2021-12-10T02:18:09Z) - Point Discriminative Learning for Unsupervised Representation Learning
on 3D Point Clouds [54.31515001741987]
We propose a point discriminative learning method for unsupervised representation learning on 3D point clouds.
We achieve this by imposing a novel point discrimination loss on the middle level and global level point features.
Our method learns powerful representations and achieves new state-of-the-art performance.
arXiv Detail & Related papers (2021-08-04T15:11:48Z) - Deep Learning for 3D Point Clouds: A Survey [58.954684611055]
This paper presents a review of recent progress in deep learning methods for point clouds.
It covers three major tasks, including 3D shape classification, 3D object detection and tracking, and 3D point cloud segmentation.
It also presents comparative results on several publicly available datasets.
arXiv Detail & Related papers (2019-12-27T09:15:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.