RBGNet: Ray-based Grouping for 3D Object Detection
- URL: http://arxiv.org/abs/2204.02251v1
- Date: Tue, 5 Apr 2022 14:42:57 GMT
- Title: RBGNet: Ray-based Grouping for 3D Object Detection
- Authors: Haiyang Wang, Shaoshuai Shi, Ze Yang, Rongyao Fang, Qi Qian, Hongsheng
Li, Bernt Schiele and Liwei Wang
- Abstract summary: We propose the RBGNet framework, a voting-based 3D detector for accurate 3D object detection from point clouds.
We propose a ray-based feature grouping module, which aggregates the point-wise features on object surfaces using a group of determined rays.
Our model achieves state-of-the-art 3D detection performance on ScanNet V2 and SUN RGB-D with remarkable performance gains.
- Score: 104.98776095895641
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As a fundamental problem in computer vision, 3D object detection is
experiencing rapid growth. To extract the point-wise features from the
irregularly and sparsely distributed points, previous methods usually take a
feature grouping module to aggregate the point features to an object candidate.
However, these methods have not yet leveraged the surface geometry of
foreground objects to enhance grouping and 3D box generation. In this paper, we
propose the RBGNet framework, a voting-based 3D detector for accurate 3D object
detection from point clouds. In order to learn better representations of object
shape to enhance cluster features for predicting 3D boxes, we propose a
ray-based feature grouping module, which aggregates the point-wise features on
object surfaces using a group of determined rays uniformly emitted from cluster
centers. Considering the fact that foreground points are more meaningful for
box estimation, we design a novel foreground biased sampling strategy in
downsample process to sample more points on object surfaces and further boost
the detection performance. Our model achieves state-of-the-art 3D detection
performance on ScanNet V2 and SUN RGB-D with remarkable performance gains. Code
will be available at https://github.com/Haiyang-W/RBGNet.
Related papers
- 3D Object Detection from Point Cloud via Voting Step Diffusion [52.9966883689137]
existing voting-based methods often receive votes from the partial surfaces of individual objects together with severe noises, leading to sub-optimal detection performance.
We propose a new method to move random 3D points toward the high-density region of the distribution by estimating the score function of the distribution with a noise conditioned score network.
Experiments on two large scale indoor 3D scene datasets, SUN RGB-D and ScanNet V2, demonstrate the superiority of our proposed method.
arXiv Detail & Related papers (2024-03-21T05:04:52Z) - SPGroup3D: Superpoint Grouping Network for Indoor 3D Object Detection [23.208654655032955]
Current 3D object detection methods for indoor scenes mainly follow the voting-and-grouping strategy to generate proposals.
We propose a novel superpoint grouping network for indoor anchor-free one-stage 3D object detection.
Experimental results demonstrate our method achieves state-of-the-art performance on ScanNet V2, SUN RGB-D, and S3DIS datasets.
arXiv Detail & Related papers (2023-12-21T08:08:02Z) - 3D Small Object Detection with Dynamic Spatial Pruning [62.72638845817799]
We propose an efficient feature pruning strategy for 3D small object detection.
We present a multi-level 3D detector named DSPDet3D which benefits from high spatial resolution.
It takes less than 2s to directly process a whole building consisting of more than 4500k points while detecting out almost all objects.
arXiv Detail & Related papers (2023-05-05T17:57:04Z) - CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds [55.44204039410225]
We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D.
Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels.
To recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module.
arXiv Detail & Related papers (2022-10-09T13:38:48Z) - A Versatile Multi-View Framework for LiDAR-based 3D Object Detection
with Guidance from Panoptic Segmentation [9.513467995188634]
3D object detection using LiDAR data is an indispensable component for autonomous driving systems.
We propose a novel multi-task framework that jointly performs 3D object detection and panoptic segmentation.
arXiv Detail & Related papers (2022-03-04T04:57:05Z) - CG-SSD: Corner Guided Single Stage 3D Object Detection from LiDAR Point
Cloud [4.110053032708927]
In a real world scene, the LiDAR can only acquire a limited object surface point clouds, but the center point of the object does not exist.
We propose a corner-guided anchor-free single-stage 3D object detection model (CG-SSD)
CG-SSD achieves the state-of-art performance on the ONCE benchmark for supervised 3D object detection using single frame point cloud data.
arXiv Detail & Related papers (2022-02-24T02:30:15Z) - SASA: Semantics-Augmented Set Abstraction for Point-based 3D Object
Detection [78.90102636266276]
We propose a novel set abstraction method named Semantics-Augmented Set Abstraction (SASA)
Based on the estimated point-wise foreground scores, we then propose a semantics-guided point sampling algorithm to help retain more important foreground points during down-sampling.
In practice, SASA shows to be effective in identifying valuable points related to foreground objects and improving feature learning for point-based 3D detection.
arXiv Detail & Related papers (2022-01-06T08:54:47Z) - Group-Free 3D Object Detection via Transformers [26.040378025818416]
We present a simple yet effective method for directly detecting 3D objects from the 3D point cloud.
Our method computes the feature of an object from all the points in the point cloud with the help of an attention mechanism in the Transformers citevaswaniattention.
With few bells and whistles, the proposed method achieves state-of-the-art 3D object detection performance on two widely used benchmarks, ScanNet V2 and SUN RGB-D.
arXiv Detail & Related papers (2021-04-01T17:59:36Z) - BirdNet+: End-to-End 3D Object Detection in LiDAR Bird's Eye View [117.44028458220427]
On-board 3D object detection in autonomous vehicles often relies on geometry information captured by LiDAR devices.
We present a fully end-to-end 3D object detection framework that can infer oriented 3D boxes solely from BEV images.
arXiv Detail & Related papers (2020-03-09T15:08:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.