CG-SSD: Corner Guided Single Stage 3D Object Detection from LiDAR Point
Cloud
- URL: http://arxiv.org/abs/2202.11868v1
- Date: Thu, 24 Feb 2022 02:30:15 GMT
- Title: CG-SSD: Corner Guided Single Stage 3D Object Detection from LiDAR Point
Cloud
- Authors: Ruiqi Ma, Chi Chen, Bisheng Yang, Deren Li, Haiping Wang, Yangzi Cong,
Zongtian Hu
- Abstract summary: In a real world scene, the LiDAR can only acquire a limited object surface point clouds, but the center point of the object does not exist.
We propose a corner-guided anchor-free single-stage 3D object detection model (CG-SSD)
CG-SSD achieves the state-of-art performance on the ONCE benchmark for supervised 3D object detection using single frame point cloud data.
- Score: 4.110053032708927
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: At present, the anchor-based or anchor-free models that use LiDAR point
clouds for 3D object detection use the center assigner strategy to infer the 3D
bounding boxes. However, in a real world scene, the LiDAR can only acquire a
limited object surface point clouds, but the center point of the object does
not exist. Obtaining the object by aggregating the incomplete surface point
clouds will bring a loss of accuracy in direction and dimension estimation. To
address this problem, we propose a corner-guided anchor-free single-stage 3D
object detection model (CG-SSD ).Firstly, 3D sparse convolution backbone
network composed of residual layers and sub-manifold sparse convolutional
layers are used to construct bird's eye view (BEV) features for further deeper
feature mining by a lite U-shaped network; Secondly, a novel corner-guided
auxiliary module (CGAM) is proposed to incorporate corner supervision signals
into the neural network. CGAM is explicitly designed and trained to detect
partially visible and invisible corners to obtains a more accurate object
feature representation, especially for small or partial occluded objects;
Finally, the deep features from both the backbone networks and CGAM module are
concatenated and fed into the head module to predict the classification and 3D
bounding boxes of the objects in the scene. The experiments demonstrate CG-SSD
achieves the state-of-art performance on the ONCE benchmark for supervised 3D
object detection using single frame point cloud data, with 62.77%mAP.
Additionally, the experiments on ONCE and Waymo Open Dataset show that CGAM can
be extended to most anchor-based models which use the BEV feature to detect
objects, as a plug-in and bring +1.17%-+14.27%AP improvement.
Related papers
- VFMM3D: Releasing the Potential of Image by Vision Foundation Model for Monocular 3D Object Detection [80.62052650370416]
monocular 3D object detection holds significant importance across various applications, including autonomous driving and robotics.
In this paper, we present VFMM3D, an innovative framework that leverages the capabilities of Vision Foundation Models (VFMs) to accurately transform single-view images into LiDAR point cloud representations.
arXiv Detail & Related papers (2024-04-15T03:12:12Z) - AOP-Net: All-in-One Perception Network for Joint LiDAR-based 3D Object
Detection and Panoptic Segmentation [9.513467995188634]
AOP-Net is a LiDAR-based multi-task framework that combines 3D object detection and panoptic segmentation.
The AOP-Net achieves state-of-the-art performance for published works on the nuScenes benchmark for both 3D object detection and panoptic segmentation tasks.
arXiv Detail & Related papers (2023-02-02T05:31:53Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - RBGNet: Ray-based Grouping for 3D Object Detection [104.98776095895641]
We propose the RBGNet framework, a voting-based 3D detector for accurate 3D object detection from point clouds.
We propose a ray-based feature grouping module, which aggregates the point-wise features on object surfaces using a group of determined rays.
Our model achieves state-of-the-art 3D detection performance on ScanNet V2 and SUN RGB-D with remarkable performance gains.
arXiv Detail & Related papers (2022-04-05T14:42:57Z) - DeepPoint: A Deep Learning Model for 3D Reconstruction in Point Clouds
via mmWave Radar [10.119506666546663]
We introduce in this paper DeepPoint, a deep learning model that generates 3D objects in point cloud format.
It takes as input the 2D depth images of an object generated by 3DRIMR's Stage 1, and outputs smooth and dense 3D point clouds of the object.
Our experiments have demonstrated that this model significantly outperforms the original 3DRIMR and other standard techniques in reconstructing 3D objects.
arXiv Detail & Related papers (2021-09-19T18:28:20Z) - Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR-based
Perception [122.53774221136193]
State-of-the-art methods for driving-scene LiDAR-based perception often project the point clouds to 2D space and then process them via 2D convolution.
A natural remedy is to utilize the 3D voxelization and 3D convolution network.
We propose a new framework for the outdoor LiDAR segmentation, where cylindrical partition and asymmetrical 3D convolution networks are designed to explore the 3D geometric pattern.
arXiv Detail & Related papers (2021-09-12T06:25:11Z) - Anchor-free 3D Single Stage Detector with Mask-Guided Attention for
Point Cloud [79.39041453836793]
We develop a novel single-stage 3D detector for point clouds in an anchor-free manner.
We overcome this by converting the voxel-based sparse 3D feature volumes into the sparse 2D feature maps.
We propose an IoU-based detection confidence re-calibration scheme to improve the correlation between the detection confidence score and the accuracy of the bounding box regression.
arXiv Detail & Related papers (2021-08-08T13:42:13Z) - Stereo RGB and Deeper LIDAR Based Network for 3D Object Detection [40.34710686994996]
3D object detection has become an emerging task in autonomous driving scenarios.
Previous works process 3D point clouds using either projection-based or voxel-based models.
We propose the Stereo RGB and Deeper LIDAR framework which can utilize semantic and spatial information simultaneously.
arXiv Detail & Related papers (2020-06-09T11:19:24Z) - LiDAR-based Online 3D Video Object Detection with Graph-based Message
Passing and Spatiotemporal Transformer Attention [100.52873557168637]
3D object detectors usually focus on the single-frame detection, while ignoring the information in consecutive point cloud frames.
In this paper, we propose an end-to-end online 3D video object detector that operates on point sequences.
arXiv Detail & Related papers (2020-04-03T06:06:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.