CenterNet3D: An Anchor Free Object Detector for Point Cloud
- URL: http://arxiv.org/abs/2007.07214v4
- Date: Mon, 25 Oct 2021 14:49:24 GMT
- Title: CenterNet3D: An Anchor Free Object Detector for Point Cloud
- Authors: Guojun Wang, Jian Wu, Bin Tian, Siyu Teng, Long Chen, Dongpu Cao
- Abstract summary: We propose an anchor-free CenterNet3D network that performs 3D object detection without anchors.
Based on the center point, we propose an anchor-free CenterNet3D network that performs 3D object detection without anchors.
Our method outperforms all state-of-the-art anchor-based one-stage methods and has comparable performance to two-stage methods as well.
- Score: 14.506796247331584
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate and fast 3D object detection from point clouds is a key task in
autonomous driving. Existing one-stage 3D object detection methods can achieve
real-time performance, however, they are dominated by anchor-based detectors
which are inefficient and require additional post-processing. In this paper, we
eliminate anchors and model an object as a single point--the center point of
its bounding box. Based on the center point, we propose an anchor-free
CenterNet3D network that performs 3D object detection without anchors. Our
CenterNet3D uses keypoint estimation to find center points and directly
regresses 3D bounding boxes. However, because inherent sparsity of point
clouds, 3D object center points are likely to be in empty space which makes it
difficult to estimate accurate boundaries. To solve this issue, we propose an
extra corner attention module to enforce the CNN backbone to pay more attention
to object boundaries. Besides, considering that one-stage detectors suffer from
the discordance between the predicted bounding boxes and corresponding
classification confidences, we develop an efficient keypoint-sensitive warping
operation to align the confidences to the predicted bounding boxes. Our
proposed CenterNet3D is non-maximum suppression free which makes it more
efficient and simpler. We evaluate CenterNet3D on the widely used KITTI dataset
and more challenging nuScenes dataset. Our method outperforms all
state-of-the-art anchor-based one-stage methods and has comparable performance
to two-stage methods as well. It has an inference speed of 20 FPS and achieves
the best speed and accuracy trade-off. Our source code will be released at
https://github.com/wangguojun2018/CenterNet3d.
Related papers
- HEDNet: A Hierarchical Encoder-Decoder Network for 3D Object Detection
in Point Clouds [19.1921315424192]
3D object detection in point clouds is important for autonomous driving systems.
A primary challenge in 3D object detection stems from the sparse distribution of points within the 3D scene.
We propose HEDNet, a hierarchical encoder-decoder network for 3D object detection.
arXiv Detail & Related papers (2023-10-31T07:32:08Z) - V-DETR: DETR with Vertex Relative Position Encoding for 3D Object
Detection [73.37781484123536]
We introduce a highly performant 3D object detector for point clouds using the DETR framework.
To address the limitation, we introduce a novel 3D Relative Position (3DV-RPE) method.
We show exceptional results on the challenging ScanNetV2 benchmark.
arXiv Detail & Related papers (2023-08-08T17:14:14Z) - 3D Small Object Detection with Dynamic Spatial Pruning [62.72638845817799]
We propose an efficient feature pruning strategy for 3D small object detection.
We present a multi-level 3D detector named DSPDet3D which benefits from high spatial resolution.
It takes less than 2s to directly process a whole building consisting of more than 4500k points while detecting out almost all objects.
arXiv Detail & Related papers (2023-05-05T17:57:04Z) - CenterNet++ for Object Detection [174.59360147041673]
Bottom-up approaches are as competitive as the top-down and enjoy higher recall.
Our approach, named CenterNet, detects each object as a triplet keypoints (top-left and bottom-right corners and the center keypoint)
On the MS-COCO dataset, CenterNet with Res2Net-101 and Swin-Transformer achieves APs of 53.7% and 57.1%, respectively.
arXiv Detail & Related papers (2022-04-18T16:45:53Z) - RBGNet: Ray-based Grouping for 3D Object Detection [104.98776095895641]
We propose the RBGNet framework, a voting-based 3D detector for accurate 3D object detection from point clouds.
We propose a ray-based feature grouping module, which aggregates the point-wise features on object surfaces using a group of determined rays.
Our model achieves state-of-the-art 3D detection performance on ScanNet V2 and SUN RGB-D with remarkable performance gains.
arXiv Detail & Related papers (2022-04-05T14:42:57Z) - Anchor-free 3D Single Stage Detector with Mask-Guided Attention for
Point Cloud [79.39041453836793]
We develop a novel single-stage 3D detector for point clouds in an anchor-free manner.
We overcome this by converting the voxel-based sparse 3D feature volumes into the sparse 2D feature maps.
We propose an IoU-based detection confidence re-calibration scheme to improve the correlation between the detection confidence score and the accuracy of the bounding box regression.
arXiv Detail & Related papers (2021-08-08T13:42:13Z) - Center-based 3D Object Detection and Tracking [8.72305226979945]
Three-dimensional objects are commonly represented as 3D boxes in a point-cloud.
This representation mimics the well-studied image-based 2D bounding-box detection but comes with additional challenges.
In this paper, we propose to represent, detect, and track 3D objects as points.
Our framework, CenterPoint, first detects centers of objects using a keypoint detector and regresses to other attributes, including 3D size, 3D orientation, and velocity.
The resulting detection and tracking algorithm is simple, efficient, and effective.
arXiv Detail & Related papers (2020-06-19T17:59:39Z) - 3D Object Detection Method Based on YOLO and K-Means for Image and Point
Clouds [1.9458156037869139]
Lidar based 3D object detection and classification tasks are essential for autonomous driving.
This paper proposes a 3D object detection method based on point cloud and image.
arXiv Detail & Related papers (2020-04-21T04:32:36Z) - PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection [76.30585706811993]
We present a novel and high-performance 3D object detection framework, named PointVoxel-RCNN (PV-RCNN)
Our proposed method deeply integrates both 3D voxel Convolutional Neural Network (CNN) and PointNet-based set abstraction.
It takes advantages of efficient learning and high-quality proposals of the 3D voxel CNN and the flexible receptive fields of the PointNet-based networks.
arXiv Detail & Related papers (2019-12-31T06:34:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.