SVGA-Net: Sparse Voxel-Graph Attention Network for 3D Object Detection
from Point Clouds
- URL: http://arxiv.org/abs/2006.04043v2
- Date: Thu, 23 Dec 2021 13:17:48 GMT
- Title: SVGA-Net: Sparse Voxel-Graph Attention Network for 3D Object Detection
from Point Clouds
- Authors: Qingdong He, Zhengning Wang, Hao Zeng, Yi Zeng, Yijun Liu
- Abstract summary: We propose Sparse Voxel-Graph Attention Network (SVGA-Net) to achieve comparable 3D detection tasks from raw LIDAR data.
SVGA-Net constructs the local complete graph within each divided 3D spherical voxel and global KNN graph through all voxels.
Experiments on KITTI detection benchmark demonstrate the efficiency of extending the graph representation to 3D object detection.
- Score: 8.906003527848636
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Accurate 3D object detection from point clouds has become a crucial component
in autonomous driving. However, the volumetric representations and the
projection methods in previous works fail to establish the relationships
between the local point sets. In this paper, we propose Sparse Voxel-Graph
Attention Network (SVGA-Net), a novel end-to-end trainable network which mainly
contains voxel-graph module and sparse-to-dense regression module to achieve
comparable 3D detection tasks from raw LIDAR data. Specifically, SVGA-Net
constructs the local complete graph within each divided 3D spherical voxel and
global KNN graph through all voxels. The local and global graphs serve as the
attention mechanism to enhance the extracted features. In addition, the novel
sparse-to-dense regression module enhances the 3D box estimation accuracy
through feature maps aggregation at different levels. Experiments on KITTI
detection benchmark demonstrate the efficiency of extending the graph
representation to 3D object detection and the proposed SVGA-Net can achieve
decent detection accuracy.
Related papers
- CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds [55.44204039410225]
We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D.
Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels.
To recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module.
arXiv Detail & Related papers (2022-10-09T13:38:48Z) - Graph R-CNN: Towards Accurate 3D Object Detection with
Semantic-Decorated Local Graph [26.226885108862735]
Two-stage detectors have gained much popularity in 3D object detection.
Most two-stage 3D detectors utilize grid points, voxel grids, or sampled keypoints for RoI feature extraction in the second stage.
This paper solves this problem in three aspects.
arXiv Detail & Related papers (2022-08-07T02:56:56Z) - RBGNet: Ray-based Grouping for 3D Object Detection [104.98776095895641]
We propose the RBGNet framework, a voting-based 3D detector for accurate 3D object detection from point clouds.
We propose a ray-based feature grouping module, which aggregates the point-wise features on object surfaces using a group of determined rays.
Our model achieves state-of-the-art 3D detection performance on ScanNet V2 and SUN RGB-D with remarkable performance gains.
arXiv Detail & Related papers (2022-04-05T14:42:57Z) - ST3D: Self-training for Unsupervised Domain Adaptation on 3D
ObjectDetection [78.71826145162092]
We present a new domain adaptive self-training pipeline, named ST3D, for unsupervised domain adaptation on 3D object detection from point clouds.
Our ST3D achieves state-of-the-art performance on all evaluated datasets and even surpasses fully supervised results on KITTI 3D object detection benchmark.
arXiv Detail & Related papers (2021-03-09T10:51:24Z) - H3D: Benchmark on Semantic Segmentation of High-Resolution 3D Point
Clouds and textured Meshes from UAV LiDAR and Multi-View-Stereo [4.263987603222371]
This paper introduces a 3D dataset which is unique in three ways.
It depicts the village of Hessigheim (Germany) henceforth referred to as H3D.
It is designed for promoting research in the field of 3D data analysis on one hand and to evaluate and rank emerging approaches.
arXiv Detail & Related papers (2021-02-10T09:33:48Z) - Voxel R-CNN: Towards High Performance Voxel-based 3D Object Detection [99.16162624992424]
We devise a simple but effective voxel-based framework, named Voxel R-CNN.
By taking full advantage of voxel features in a two stage approach, our method achieves comparable detection accuracy with state-of-the-art point-based models.
Our results show that Voxel R-CNN delivers a higher detection accuracy while maintaining a realtime frame processing rate, emphi.e, at a speed of 25 FPS on an NVIDIA 2080 Ti GPU.
arXiv Detail & Related papers (2020-12-31T17:02:46Z) - Dynamic Edge Weights in Graph Neural Networks for 3D Object Detection [0.0]
We propose an attention based feature aggregation technique in graph neural network (GNN) for detecting objects in LiDAR scan.
In each layer of the GNN, apart from the linear transformation which maps the per node input features to the corresponding higher level features, a per node masked attention is also performed.
The experiments on KITTI dataset show that our method yields comparable results for 3D object detection.
arXiv Detail & Related papers (2020-09-17T12:56:17Z) - InfoFocus: 3D Object Detection for Autonomous Driving with Dynamic
Information Modeling [65.47126868838836]
We propose a novel 3D object detection framework with dynamic information modeling.
Coarse predictions are generated in the first stage via a voxel-based region proposal network.
Experiments are conducted on the large-scale nuScenes 3D detection benchmark.
arXiv Detail & Related papers (2020-07-16T18:27:08Z) - Local Grid Rendering Networks for 3D Object Detection in Point Clouds [98.02655863113154]
CNNs are powerful but it would be computationally costly to directly apply convolutions on point data after voxelizing the entire point clouds to a dense regular 3D grid.
We propose a novel and principled Local Grid Rendering (LGR) operation to render the small neighborhood of a subset of input points into a low-resolution 3D grid independently.
We validate LGR-Net for 3D object detection on the challenging ScanNet and SUN RGB-D datasets.
arXiv Detail & Related papers (2020-07-04T13:57:43Z) - Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud [3.04585143845864]
We propose a graph neural network to detect objects from a LiDAR point cloud.
We encode the point cloud efficiently in a fixed radius near-neighbors graph.
In Point-GNN, we propose an auto-registration mechanism to reduce translation variance.
arXiv Detail & Related papers (2020-03-02T23:44:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.