Prototypical VoteNet for Few-Shot 3D Point Cloud Object Detection
- URL: http://arxiv.org/abs/2210.05593v1
- Date: Tue, 11 Oct 2022 16:25:38 GMT
- Title: Prototypical VoteNet for Few-Shot 3D Point Cloud Object Detection
- Authors: Shizhen Zhao, Xiaojuan Qi
- Abstract summary: Prototypical VoteNet is a few-shot 3D point cloud object detection approach.
It incorporates two new modules: Prototypical Vote Module (PVM) and Prototypical Head Module (PHM)
- Score: 37.48935478836176
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most existing 3D point cloud object detection approaches heavily rely on
large amounts of labeled training data. However, the labeling process is costly
and time-consuming. This paper considers few-shot 3D point cloud object
detection, where only a few annotated samples of novel classes are needed with
abundant samples of base classes. To this end, we propose Prototypical VoteNet
to recognize and localize novel instances, which incorporates two new modules:
Prototypical Vote Module (PVM) and Prototypical Head Module (PHM).
Specifically, as the 3D basic geometric structures can be shared among
categories, PVM is designed to leverage class-agnostic geometric prototypes,
which are learned from base classes, to refine local features of novel
categories.Then PHM is proposed to utilize class prototypes to enhance the
global feature of each object, facilitating subsequent object localization and
classification, which is trained by the episodic training strategy. To evaluate
the model in this new setting, we contribute two new benchmark datasets,
FS-ScanNet and FS-SUNRGBD. We conduct extensive experiments to demonstrate the
effectiveness of Prototypical VoteNet, and our proposed method shows
significant and consistent improvements compared to baselines on two benchmark
datasets.
Related papers
- CP-VoteNet: Contrastive Prototypical VoteNet for Few-Shot Point Cloud Object Detection [7.205000222081269]
Few-shot point cloud 3D object detection (FS3D) aims to identify and localise objects of novel classes from point clouds.
We introduce contrastive semantics mining, which enables the network to extract discriminative categorical features.
Through refined primitive geometric structures, the transferability of feature encoding from base to novel classes is significantly enhanced.
arXiv Detail & Related papers (2024-08-30T06:13:49Z) - Rethinking Few-shot 3D Point Cloud Semantic Segmentation [62.80639841429669]
This paper revisits few-shot 3D point cloud semantic segmentation (FS-PCS)
We focus on two significant issues in the state-of-the-art: foreground leakage and sparse point distribution.
To address these issues, we introduce a standardized FS-PCS setting, upon which a new benchmark is built.
arXiv Detail & Related papers (2024-03-01T15:14:47Z) - Dynamic Clustering Transformer Network for Point Cloud Segmentation [23.149220817575195]
We propose a novel 3D point cloud representation network, called Dynamic Clustering Transformer Network (DCTNet)
It has an encoder-decoder architecture, allowing for both local and global feature learning.
Our method was evaluated on an object-based dataset (ShapeNet), an urban navigation dataset (Toronto-3D), and a multispectral LiDAR dataset.
arXiv Detail & Related papers (2023-05-30T01:11:05Z) - Weakly Supervised 3D Point Cloud Segmentation via Multi-Prototype
Learning [37.76664203157892]
A fundamental challenge here lies in the large intra-class variations of local geometric structure, resulting in subclasses within a semantic class.
We leverage this intuition and opt for maintaining an individual classifier for each subclass.
Our hypothesis is also verified given the consistent discovery of semantic subclasses at no cost of additional annotations.
arXiv Detail & Related papers (2022-05-06T11:07:36Z) - RBGNet: Ray-based Grouping for 3D Object Detection [104.98776095895641]
We propose the RBGNet framework, a voting-based 3D detector for accurate 3D object detection from point clouds.
We propose a ray-based feature grouping module, which aggregates the point-wise features on object surfaces using a group of determined rays.
Our model achieves state-of-the-art 3D detection performance on ScanNet V2 and SUN RGB-D with remarkable performance gains.
arXiv Detail & Related papers (2022-04-05T14:42:57Z) - Learn to Learn Metric Space for Few-Shot Segmentation of 3D Shapes [17.217954254022573]
We introduce a meta-learning-based method for few-shot 3D shape segmentation where only a few labeled samples are provided for the unseen classes.
We demonstrate the superior performance of our proposed on the ShapeNet part dataset under the few-shot scenario, compared with well-established baseline and state-of-the-art semi-supervised methods.
arXiv Detail & Related papers (2021-07-07T01:47:00Z) - UniT: Unified Knowledge Transfer for Any-shot Object Detection and
Segmentation [52.487469544343305]
Methods for object detection and segmentation rely on large scale instance-level annotations for training.
We propose an intuitive and unified semi-supervised model that is applicable to a range of supervision.
arXiv Detail & Related papers (2020-06-12T22:45:47Z) - Fine-Grained 3D Shape Classification with Hierarchical Part-View
Attentions [70.0171362989609]
We propose a novel fine-grained 3D shape classification method named FG3D-Net to capture the fine-grained local details of 3D shapes from multiple rendered views.
Our results under the fine-grained 3D shape dataset show that our method outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2020-05-26T06:53:19Z) - SSN: Shape Signature Networks for Multi-class Object Detection from
Point Clouds [96.51884187479585]
We propose a novel 3D shape signature to explore the shape information from point clouds.
By incorporating operations of symmetry, convex hull and chebyshev fitting, the proposed shape sig-nature is not only compact and effective but also robust to the noise.
Experiments show that the proposed method performs remarkably better than existing methods on two large-scale datasets.
arXiv Detail & Related papers (2020-04-06T16:01:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.