Graph R-CNN: Towards Accurate 3D Object Detection with
Semantic-Decorated Local Graph
- URL: http://arxiv.org/abs/2208.03624v1
- Date: Sun, 7 Aug 2022 02:56:56 GMT
- Title: Graph R-CNN: Towards Accurate 3D Object Detection with
Semantic-Decorated Local Graph
- Authors: Honghui Yang, Zili Liu, Xiaopei Wu, Wenxiao Wang, Wei Qian, Xiaofei
He, Deng Cai
- Abstract summary: Two-stage detectors have gained much popularity in 3D object detection.
Most two-stage 3D detectors utilize grid points, voxel grids, or sampled keypoints for RoI feature extraction in the second stage.
This paper solves this problem in three aspects.
- Score: 26.226885108862735
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Two-stage detectors have gained much popularity in 3D object detection. Most
two-stage 3D detectors utilize grid points, voxel grids, or sampled keypoints
for RoI feature extraction in the second stage. Such methods, however, are
inefficient in handling unevenly distributed and sparse outdoor points. This
paper solves this problem in three aspects. 1) Dynamic Point Aggregation. We
propose the patch search to quickly search points in a local region for each 3D
proposal. The dynamic farthest voxel sampling is then applied to evenly sample
the points. Especially, the voxel size varies along the distance to accommodate
the uneven distribution of points. 2) RoI-graph Pooling. We build local graphs
on the sampled points to better model contextual information and mine point
relations through iterative message passing. 3) Visual Features Augmentation.
We introduce a simple yet effective fusion strategy to compensate for sparse
LiDAR points with limited semantic cues. Based on these modules, we construct
our Graph R-CNN as the second stage, which can be applied to existing one-stage
detectors to consistently improve the detection performance. Extensive
experiments show that Graph R-CNN outperforms the state-of-the-art 3D detection
models by a large margin on both the KITTI and Waymo Open Dataset. And we rank
first place on the KITTI BEV car detection leaderboard. Code will be available
at \url{https://github.com/Nightmare-n/GraphRCNN}.
Related papers
- V-DETR: DETR with Vertex Relative Position Encoding for 3D Object
Detection [73.37781484123536]
We introduce a highly performant 3D object detector for point clouds using the DETR framework.
To address the limitation, we introduce a novel 3D Relative Position (3DV-RPE) method.
We show exceptional results on the challenging ScanNetV2 benchmark.
arXiv Detail & Related papers (2023-08-08T17:14:14Z) - CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds [55.44204039410225]
We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D.
Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels.
To recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module.
arXiv Detail & Related papers (2022-10-09T13:38:48Z) - LiDAR R-CNN: An Efficient and Universal 3D Object Detector [20.17906188581305]
LiDAR-based 3D detection in point cloud is essential in the perception system of autonomous driving.
We present LiDAR R-CNN, a second stage detector that can generally improve any existing 3D detector.
In particular, based on one variant of PointPillars, our method could achieve new state-of-the-art results with minor cost.
arXiv Detail & Related papers (2021-03-29T03:01:21Z) - PV-RCNN++: Point-Voxel Feature Set Abstraction With Local Vector
Representation for 3D Object Detection [100.60209139039472]
We propose the PointVoxel Region based Convolution Neural Networks (PVRCNNs) for accurate 3D detection from point clouds.
Our proposed PV-RCNNs significantly outperform previous state-of-the-art 3D detection methods on both the Open dataset and the highly-competitive KITTI benchmark.
arXiv Detail & Related papers (2021-01-31T14:51:49Z) - Voxel R-CNN: Towards High Performance Voxel-based 3D Object Detection [99.16162624992424]
We devise a simple but effective voxel-based framework, named Voxel R-CNN.
By taking full advantage of voxel features in a two stage approach, our method achieves comparable detection accuracy with state-of-the-art point-based models.
Our results show that Voxel R-CNN delivers a higher detection accuracy while maintaining a realtime frame processing rate, emphi.e, at a speed of 25 FPS on an NVIDIA 2080 Ti GPU.
arXiv Detail & Related papers (2020-12-31T17:02:46Z) - SVGA-Net: Sparse Voxel-Graph Attention Network for 3D Object Detection
from Point Clouds [8.906003527848636]
We propose Sparse Voxel-Graph Attention Network (SVGA-Net) to achieve comparable 3D detection tasks from raw LIDAR data.
SVGA-Net constructs the local complete graph within each divided 3D spherical voxel and global KNN graph through all voxels.
Experiments on KITTI detection benchmark demonstrate the efficiency of extending the graph representation to 3D object detection.
arXiv Detail & Related papers (2020-06-07T05:01:06Z) - D3Feat: Joint Learning of Dense Detection and Description of 3D Local
Features [51.04841465193678]
We leverage a 3D fully convolutional network for 3D point clouds.
We propose a novel and practical learning mechanism that densely predicts both a detection score and a description feature for each 3D point.
Our method achieves state-of-the-art results in both indoor and outdoor scenarios.
arXiv Detail & Related papers (2020-03-06T12:51:09Z) - Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud [3.04585143845864]
We propose a graph neural network to detect objects from a LiDAR point cloud.
We encode the point cloud efficiently in a fixed radius near-neighbors graph.
In Point-GNN, we propose an auto-registration mechanism to reduce translation variance.
arXiv Detail & Related papers (2020-03-02T23:44:12Z) - PV-RCNN: Point-Voxel Feature Set Abstraction for 3D Object Detection [76.30585706811993]
We present a novel and high-performance 3D object detection framework, named PointVoxel-RCNN (PV-RCNN)
Our proposed method deeply integrates both 3D voxel Convolutional Neural Network (CNN) and PointNet-based set abstraction.
It takes advantages of efficient learning and high-quality proposals of the 3D voxel CNN and the flexible receptive fields of the PointNet-based networks.
arXiv Detail & Related papers (2019-12-31T06:34:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.