Towards Open World Active Learning for 3D Object Detection
- URL: http://arxiv.org/abs/2310.10391v1
- Date: Mon, 16 Oct 2023 13:32:53 GMT
- Title: Towards Open World Active Learning for 3D Object Detection
- Authors: Zhuoxiao Chen, Yadan Luo, Zixin Wang, Zijian Wang, Xin Yu, Zi Huang
- Abstract summary: We introduce Open World Active Learning for 3D Object Detection (OWAL-3D)
OWAL-3D aims at selecting a small number of 3D boxes to annotate while maximizing detection performance on both known and unknown classes.
We unify both relational constraints into a simple and effective AL strategy namely OpenCRB, which guides to acquisition of informative point clouds.
- Score: 43.242426340854905
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Significant strides have been made in closed world 3D object detection,
testing systems in environments with known classes. However, the challenge
arises in open world scenarios where new object classes appear. Existing
efforts sequentially learn novel classes from streams of labeled data at a
significant annotation cost, impeding efficient deployment to the wild. To seek
effective solutions, we investigate a more practical yet challenging research
task: Open World Active Learning for 3D Object Detection (OWAL-3D), aiming at
selecting a small number of 3D boxes to annotate while maximizing detection
performance on both known and unknown classes. The core difficulty centers on
striking a balance between mining more unknown instances and minimizing the
labeling expenses of point clouds. Empirically, our study finds the harmonious
and inverse relationship between box quantities and their confidences can help
alleviate the dilemma, avoiding the repeated selection of common known
instances and focusing on uncertain objects that are potentially unknown. We
unify both relational constraints into a simple and effective AL strategy
namely OpenCRB, which guides to acquisition of informative point clouds with
the least amount of boxes to label. Furthermore, we develop a comprehensive
codebase for easy reproducing and future research, supporting 15 baseline
methods (i.e., active learning, out-of-distribution detection and open world
detection), 2 types of modern 3D detectors (i.e., one-stage SECOND and
two-stage PV-RCNN) and 3 benchmark 3D datasets (i.e., KITTI, nuScenes and
Waymo). Extensive experiments evidence that the proposed Open-CRB demonstrates
superiority and flexibility in recognizing both novel and shared categories
with very limited labeling costs, compared to state-of-the-art baselines.
Related papers
- OC3D: Weakly Supervised Outdoor 3D Object Detection with Only Coarse Click Annotation [21.28388955131355]
Training 3D detectors from the LiDAR point cloud typically relies on expensive bounding box annotations.
This paper presents OC3D, an innovative weakly supervised method requiring only coarse clicks on the bird's eye view of the 3D point cloud.
arXiv Detail & Related papers (2024-08-15T11:34:53Z) - Find n' Propagate: Open-Vocabulary 3D Object Detection in Urban Environments [67.83787474506073]
We tackle the limitations of current LiDAR-based 3D object detection systems.
We introduce a universal textscFind n' Propagate approach for 3D OV tasks.
We achieve up to a 3.97-fold increase in Average Precision (AP) for novel object classes.
arXiv Detail & Related papers (2024-03-20T12:51:30Z) - Open-Pose 3D Zero-Shot Learning: Benchmark and Challenges [23.663199578392447]
We propose a more realistic and challenging scenario named open-pose 3D zero-shot classification.
First, we revisit the current research on 3D zero-shot classification.
We propose two benchmark datasets specifically designed for the open-pose setting.
arXiv Detail & Related papers (2023-12-12T07:52:33Z) - Exploring Active 3D Object Detection from a Generalization Perspective [58.597942380989245]
Uncertainty-based active learning policies fail to balance the trade-off between point cloud informativeness and box-level annotation costs.
We propose textscCrb, which hierarchically filters out the point clouds of redundant 3D bounding box labels.
Experiments show that the proposed approach outperforms existing active learning strategies.
arXiv Detail & Related papers (2023-01-23T02:43:03Z) - Towards Open Set 3D Learning: A Benchmark on Object Point Clouds [17.145309633743747]
This paper provides the first broad study on Open Set 3D learning.
We introduce a novel testbed with settings of increasing difficulty in terms of category semantic shift.
We investigate the related out-of-distribution and Open Set 2D literature to understand if and how their most recent approaches are effective on 3D data.
arXiv Detail & Related papers (2022-07-23T17:00:45Z) - Exploring Diversity-based Active Learning for 3D Object Detection in Autonomous Driving [45.405303803618]
We investigate diversity-based active learning (AL) as a potential solution to alleviate the annotation burden.
We propose a novel acquisition function that enforces spatial and temporal diversity in the selected samples.
We demonstrate the effectiveness of the proposed method on the nuScenes dataset and show that it outperforms existing AL strategies significantly.
arXiv Detail & Related papers (2022-05-16T14:21:30Z) - From Multi-View to Hollow-3D: Hallucinated Hollow-3D R-CNN for 3D Object
Detection [101.20784125067559]
We propose a new architecture, namely Hallucinated Hollow-3D R-CNN, to address the problem of 3D object detection.
In our approach, we first extract the multi-view features by sequentially projecting the point clouds into the perspective view and the bird-eye view.
The 3D objects are detected via a box refinement module with a novel Hierarchical Voxel RoI Pooling operation.
arXiv Detail & Related papers (2021-07-30T02:00:06Z) - D3Feat: Joint Learning of Dense Detection and Description of 3D Local
Features [51.04841465193678]
We leverage a 3D fully convolutional network for 3D point clouds.
We propose a novel and practical learning mechanism that densely predicts both a detection score and a description feature for each 3D point.
Our method achieves state-of-the-art results in both indoor and outdoor scenarios.
arXiv Detail & Related papers (2020-03-06T12:51:09Z) - SESS: Self-Ensembling Semi-Supervised 3D Object Detection [138.80825169240302]
We propose SESS, a self-ensembling semi-supervised 3D object detection framework. Specifically, we design a thorough perturbation scheme to enhance generalization of the network on unlabeled and new unseen data.
Our SESS achieves competitive performance compared to the state-of-the-art fully-supervised method by using only 50% labeled data.
arXiv Detail & Related papers (2019-12-26T08:48:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.