Not All Points Are Equal: Learning Highly Efficient Point-based
Detectors for 3D LiDAR Point Clouds
- URL: http://arxiv.org/abs/2203.11139v1
- Date: Mon, 21 Mar 2022 17:14:02 GMT
- Title: Not All Points Are Equal: Learning Highly Efficient Point-based
Detectors for 3D LiDAR Point Clouds
- Authors: Yifan Zhang, Qingyong Hu, Guoquan Xu, Yanxin Ma, Jianwei Wan, Yulan
Guo
- Abstract summary: We propose a highly-efficient single-stage point-based 3D detector called IA-SSD.
We exploit two learnable, task-oriented, instance-aware downsampling strategies to hierarchically select the foreground points.
Experiments conducted on several large-scale detection benchmarks demonstrate the competitive performance of our IA-SSD.
- Score: 29.762645632148097
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: We study the problem of efficient object detection of 3D LiDAR point clouds.
To reduce the memory and computational cost, existing point-based pipelines
usually adopt task-agnostic random sampling or farthest point sampling to
progressively downsample input point clouds, despite the fact that not all
points are equally important to the task of object detection. In particular,
the foreground points are inherently more important than background points for
object detectors. Motivated by this, we propose a highly-efficient single-stage
point-based 3D detector in this paper, termed IA-SSD. The key of our approach
is to exploit two learnable, task-oriented, instance-aware downsampling
strategies to hierarchically select the foreground points belonging to objects
of interest. Additionally, we also introduce a contextual centroid perception
module to further estimate precise instance centers. Finally, we build our
IA-SSD following the encoder-only architecture for efficiency. Extensive
experiments conducted on several large-scale detection benchmarks demonstrate
the competitive performance of our IA-SSD. Thanks to the low memory footprint
and a high degree of parallelism, it achieves a superior speed of 80+
frames-per-second on the KITTI dataset with a single RTX2080Ti GPU. The code is
available at \url{https://github.com/yifanzhang713/IA-SSD}.
Related papers
- V-DETR: DETR with Vertex Relative Position Encoding for 3D Object
Detection [73.37781484123536]
We introduce a highly performant 3D object detector for point clouds using the DETR framework.
To address the limitation, we introduce a novel 3D Relative Position (3DV-RPE) method.
We show exceptional results on the challenging ScanNetV2 benchmark.
arXiv Detail & Related papers (2023-08-08T17:14:14Z) - 3D Small Object Detection with Dynamic Spatial Pruning [62.72638845817799]
We propose an efficient feature pruning strategy for 3D small object detection.
We present a multi-level 3D detector named DSPDet3D which benefits from high spatial resolution.
It takes less than 2s to directly process a whole building consisting of more than 4500k points while detecting out almost all objects.
arXiv Detail & Related papers (2023-05-05T17:57:04Z) - 3D Cascade RCNN: High Quality Object Detection in Point Clouds [122.42455210196262]
We present 3D Cascade RCNN, which allocates multiple detectors based on the voxelized point clouds in a cascade paradigm.
We validate the superiority of our proposed 3D Cascade RCNN, when comparing to state-of-the-art 3D object detection techniques.
arXiv Detail & Related papers (2022-11-15T15:58:36Z) - PSA-Det3D: Pillar Set Abstraction for 3D object Detection [14.788139868324155]
We propose a pillar set abstraction (PSA) and foreground point compensation (FPC) to improve the detection performance for small object.
The experiments on the KITTI 3D detection benchmark show that our proposed PSA-Det3D outperforms other algorithms with high accuracy for small object detection.
arXiv Detail & Related papers (2022-10-20T03:05:34Z) - SASA: Semantics-Augmented Set Abstraction for Point-based 3D Object
Detection [78.90102636266276]
We propose a novel set abstraction method named Semantics-Augmented Set Abstraction (SASA)
Based on the estimated point-wise foreground scores, we then propose a semantics-guided point sampling algorithm to help retain more important foreground points during down-sampling.
In practice, SASA shows to be effective in identifying valuable points related to foreground objects and improving feature learning for point-based 3D detection.
arXiv Detail & Related papers (2022-01-06T08:54:47Z) - Learning Semantic Segmentation of Large-Scale Point Clouds with Random
Sampling [52.464516118826765]
We introduce RandLA-Net, an efficient and lightweight neural architecture to infer per-point semantics for large-scale point clouds.
The key to our approach is to use random point sampling instead of more complex point selection approaches.
Our RandLA-Net can process 1 million points in a single pass up to 200x faster than existing approaches.
arXiv Detail & Related papers (2021-07-06T05:08:34Z) - Boundary-Aware Dense Feature Indicator for Single-Stage 3D Object
Detection from Point Clouds [32.916690488130506]
We propose a universal module that helps 3D detectors focus on the densest region of the point clouds in a boundary-aware manner.
Experiments on KITTI dataset show that DENFI improves the performance of the baseline single-stage detector remarkably.
arXiv Detail & Related papers (2020-04-01T01:21:23Z) - 3DSSD: Point-based 3D Single Stage Object Detector [61.67928229961813]
We present a point-based 3D single stage object detector, named 3DSSD, achieving a good balance between accuracy and efficiency.
Our method outperforms all state-of-the-art voxel-based single stage methods by a large margin, and has comparable performance to two stage point-based methods as well.
arXiv Detail & Related papers (2020-02-24T12:01:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.