Towards Open Set 3D Learning: A Benchmark on Object Point Clouds
- URL: http://arxiv.org/abs/2207.11554v1
- Date: Sat, 23 Jul 2022 17:00:45 GMT
- Title: Towards Open Set 3D Learning: A Benchmark on Object Point Clouds
- Authors: Antonio Alliegro, Francesco Cappio Borlino, Tatiana Tommasi
- Abstract summary: This paper provides the first broad study on Open Set 3D learning.
We introduce a novel testbed with settings of increasing difficulty in terms of category semantic shift.
We investigate the related out-of-distribution and Open Set 2D literature to understand if and how their most recent approaches are effective on 3D data.
- Score: 17.145309633743747
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the last years, there has been significant progress in the field of 3D
learning on classification, detection and segmentation problems. The vast
majority of the existing studies focus on canonical closed-set conditions,
neglecting the intrinsic open nature of the real-world. This limits the
abilities of autonomous systems involved in safety-critical applications that
require managing novel and unknown signals. In this context exploiting 3D data
can be a valuable asset since it conveys rich information about the geometry of
sensed objects and scenes. This paper provides the first broad study on Open
Set 3D learning. We introduce a novel testbed with settings of increasing
difficulty in terms of category semantic shift and cover both in-domain
(synthetic-to-synthetic) and cross-domain (synthetic-to-real) scenarios.
Moreover, we investigate the related out-of-distribution and Open Set 2D
literature to understand if and how their most recent approaches are effective
on 3D data. Our extensive benchmark positions several algorithms in the same
coherent picture, revealing their strengths and limitations. The results of our
analysis may serve as a reliable foothold for future tailored Open Set 3D
models.
Related papers
- Open3DTrack: Towards Open-Vocabulary 3D Multi-Object Tracking [73.05477052645885]
We introduce open-vocabulary 3D tracking, which extends the scope of 3D tracking to include objects beyond predefined categories.
We propose a novel approach that integrates open-vocabulary capabilities into a 3D tracking framework, allowing for generalization to unseen object classes.
arXiv Detail & Related papers (2024-10-02T15:48:42Z) - OpenSU3D: Open World 3D Scene Understanding using Foundation Models [2.1262749936758216]
We present a novel, scalable approach for constructing open set, instance-level 3D scene representations.
Existing methods require pre-constructed 3D scenes and face scalability issues due to per-point feature vector learning.
We evaluate our proposed approach on multiple scenes from ScanNet and Replica datasets demonstrating zero-shot generalization capabilities.
arXiv Detail & Related papers (2024-07-19T13:01:12Z) - MMScan: A Multi-Modal 3D Scene Dataset with Hierarchical Grounded Language Annotations [55.022519020409405]
This paper builds the first largest ever multi-modal 3D scene dataset and benchmark with hierarchical grounded language annotations, MMScan.
The resulting multi-modal 3D dataset encompasses 1.4M meta-annotated captions on 109k objects and 7.7k regions as well as over 3.04M diverse samples for 3D visual grounding and question-answering benchmarks.
arXiv Detail & Related papers (2024-06-13T17:59:30Z) - OV-Uni3DETR: Towards Unified Open-Vocabulary 3D Object Detection via Cycle-Modality Propagation [67.56268991234371]
OV-Uni3DETR achieves the state-of-the-art performance on various scenarios, surpassing existing methods by more than 6% on average.
Code and pre-trained models will be released later.
arXiv Detail & Related papers (2024-03-28T17:05:04Z) - Find n' Propagate: Open-Vocabulary 3D Object Detection in Urban Environments [67.83787474506073]
We tackle the limitations of current LiDAR-based 3D object detection systems.
We introduce a universal textscFind n' Propagate approach for 3D OV tasks.
We achieve up to a 3.97-fold increase in Average Precision (AP) for novel object classes.
arXiv Detail & Related papers (2024-03-20T12:51:30Z) - Open-Pose 3D Zero-Shot Learning: Benchmark and Challenges [23.663199578392447]
We propose a more realistic and challenging scenario named open-pose 3D zero-shot classification.
First, we revisit the current research on 3D zero-shot classification.
We propose two benchmark datasets specifically designed for the open-pose setting.
arXiv Detail & Related papers (2023-12-12T07:52:33Z) - Open-CRB: Towards Open World Active Learning for 3D Object Detection [40.80953254074535]
LiDAR-based 3D object detection has recently seen significant advancements through active learning (AL)
In real-world deployments where streaming point clouds may include unknown or novel objects, the ability of current AL methods to capture such objects remains unexplored.
This paper investigates a more practical and challenging research task: Open World Active Learning for 3D Object Detection (OWAL-3D)
arXiv Detail & Related papers (2023-10-16T13:32:53Z) - OpenPatch: a 3D patchwork for Out-Of-Distribution detection [16.262921993755892]
We present an approach for the task of semantic novelty detection on real-world point cloud samples when the reference known data are synthetic.
OpenPatch builds on a large pre-trained model and simply extracts from its intermediate features a set of patch representations that describe each known class.
We demonstrate that OpenPatch excels in both the full and few-shot known sample scenarios.
arXiv Detail & Related papers (2023-10-05T08:49:51Z) - CMR3D: Contextualized Multi-Stage Refinement for 3D Object Detection [57.44434974289945]
We propose Contextualized Multi-Stage Refinement for 3D Object Detection (CMR3D) framework.
Our framework takes a 3D scene as input and strives to explicitly integrate useful contextual information of the scene.
In addition to 3D object detection, we investigate the effectiveness of our framework for the problem of 3D object counting.
arXiv Detail & Related papers (2022-09-13T05:26:09Z) - RandomRooms: Unsupervised Pre-training from Synthetic Shapes and
Randomized Layouts for 3D Object Detection [138.2892824662943]
A promising solution is to make better use of the synthetic dataset, which consists of CAD object models, to boost the learning on real datasets.
Recent work on 3D pre-training exhibits failure when transfer features learned on synthetic objects to other real-world applications.
In this work, we put forward a new method called RandomRooms to accomplish this objective.
arXiv Detail & Related papers (2021-08-17T17:56:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.