SPGroup3D: Superpoint Grouping Network for Indoor 3D Object Detection
- URL: http://arxiv.org/abs/2312.13641v1
- Date: Thu, 21 Dec 2023 08:08:02 GMT
- Title: SPGroup3D: Superpoint Grouping Network for Indoor 3D Object Detection
- Authors: Yun Zhu, Le Hui, Yaqi Shen, Jin Xie
- Abstract summary: Current 3D object detection methods for indoor scenes mainly follow the voting-and-grouping strategy to generate proposals.
We propose a novel superpoint grouping network for indoor anchor-free one-stage 3D object detection.
Experimental results demonstrate our method achieves state-of-the-art performance on ScanNet V2, SUN RGB-D, and S3DIS datasets.
- Score: 23.208654655032955
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current 3D object detection methods for indoor scenes mainly follow the
voting-and-grouping strategy to generate proposals. However, most methods
utilize instance-agnostic groupings, such as ball query, leading to
inconsistent semantic information and inaccurate regression of the proposals.
To this end, we propose a novel superpoint grouping network for indoor
anchor-free one-stage 3D object detection. Specifically, we first adopt an
unsupervised manner to partition raw point clouds into superpoints, areas with
semantic consistency and spatial similarity. Then, we design a geometry-aware
voting module that adapts to the centerness in anchor-free detection by
constraining the spatial relationship between superpoints and object centers.
Next, we present a superpoint-based grouping module to explore the consistent
representation within proposals. This module includes a superpoint attention
layer to learn feature interaction between neighboring superpoints, and a
superpoint-voxel fusion layer to propagate the superpoint-level information to
the voxel level. Finally, we employ effective multiple matching to capitalize
on the dynamic receptive fields of proposals based on superpoints during the
training. Experimental results demonstrate our method achieves state-of-the-art
performance on ScanNet V2, SUN RGB-D, and S3DIS datasets in the indoor
one-stage 3D object detection. Source code is available at
https://github.com/zyrant/SPGroup3D.
Related papers
- 3D Object Detection from Point Cloud via Voting Step Diffusion [52.9966883689137]
existing voting-based methods often receive votes from the partial surfaces of individual objects together with severe noises, leading to sub-optimal detection performance.
We propose a new method to move random 3D points toward the high-density region of the distribution by estimating the score function of the distribution with a noise conditioned score network.
Experiments on two large scale indoor 3D scene datasets, SUN RGB-D and ScanNet V2, demonstrate the superiority of our proposed method.
arXiv Detail & Related papers (2024-03-21T05:04:52Z) - Open3DIS: Open-Vocabulary 3D Instance Segmentation with 2D Mask Guidance [49.14140194332482]
We introduce Open3DIS, a novel solution designed to tackle the problem of Open-Vocabulary Instance within 3D scenes.
Objects within 3D environments exhibit diverse shapes, scales, and colors, making precise instance-level identification a challenging task.
arXiv Detail & Related papers (2023-12-17T10:07:03Z) - CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds [55.44204039410225]
We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D.
Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels.
To recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module.
arXiv Detail & Related papers (2022-10-09T13:38:48Z) - RBGNet: Ray-based Grouping for 3D Object Detection [104.98776095895641]
We propose the RBGNet framework, a voting-based 3D detector for accurate 3D object detection from point clouds.
We propose a ray-based feature grouping module, which aggregates the point-wise features on object surfaces using a group of determined rays.
Our model achieves state-of-the-art 3D detection performance on ScanNet V2 and SUN RGB-D with remarkable performance gains.
arXiv Detail & Related papers (2022-04-05T14:42:57Z) - SASA: Semantics-Augmented Set Abstraction for Point-based 3D Object
Detection [78.90102636266276]
We propose a novel set abstraction method named Semantics-Augmented Set Abstraction (SASA)
Based on the estimated point-wise foreground scores, we then propose a semantics-guided point sampling algorithm to help retain more important foreground points during down-sampling.
In practice, SASA shows to be effective in identifying valuable points related to foreground objects and improving feature learning for point-based 3D detection.
arXiv Detail & Related papers (2022-01-06T08:54:47Z) - 3D Object Detection Combining Semantic and Geometric Features from Point
Clouds [19.127930862527666]
We propose a novel end-to-end two-stage 3D object detector named SGNet for point clouds scenes.
The VTPM is a Voxel-Point-Based Module that finally implements 3D object detection in point space.
As of September 19, 2021, for KITTI dataset, SGNet ranked 1st in 3D and BEV detection on cyclists with easy difficulty level, and 2nd in the 3D detection of moderate cyclists.
arXiv Detail & Related papers (2021-10-10T04:43:27Z) - Instance Segmentation in 3D Scenes using Semantic Superpoint Tree
Networks [64.27814530457042]
We propose an end-to-end solution of Semantic Superpoint Tree Network (SSTNet) for proposing object instances from scene points.
Key in SSTNet is an intermediate, semantic superpoint tree (SST), which is constructed based on the learned semantic features of superpoints.
SSTNet ranks top on the ScanNet (V2) leaderboard, with 2% higher of mAP than the second best method.
arXiv Detail & Related papers (2021-08-17T07:25:14Z) - VIN: Voxel-based Implicit Network for Joint 3D Object Detection and
Segmentation for Lidars [12.343333815270402]
A unified neural network structure is presented for joint 3D object detection and point cloud segmentation.
We leverage rich supervision from both detection and segmentation labels rather than using just one of them.
arXiv Detail & Related papers (2021-07-07T02:16:20Z) - Group-Free 3D Object Detection via Transformers [26.040378025818416]
We present a simple yet effective method for directly detecting 3D objects from the 3D point cloud.
Our method computes the feature of an object from all the points in the point cloud with the help of an attention mechanism in the Transformers citevaswaniattention.
With few bells and whistles, the proposed method achieves state-of-the-art 3D object detection performance on two widely used benchmarks, ScanNet V2 and SUN RGB-D.
arXiv Detail & Related papers (2021-04-01T17:59:36Z) - 3D-MPA: Multi Proposal Aggregation for 3D Semantic Instance Segmentation [26.169985423367393]
3D-MPA is a method for instance segmentation on 3D point clouds.
We learn proposal features from grouped point features that voted for the same object center.
A graph convolutional network introduces inter-proposal relations.
arXiv Detail & Related papers (2020-03-30T23:28:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.