3DLG-Detector: 3D Object Detection via Simultaneous Local-Global Feature
Learning
- URL: http://arxiv.org/abs/2208.14796v1
- Date: Wed, 31 Aug 2022 12:23:40 GMT
- Title: 3DLG-Detector: 3D Object Detection via Simultaneous Local-Global Feature
Learning
- Authors: Baian Chen, Liangliang Nan, Haoran Xie, Dening Lu, Fu Lee Wang and
Mingqiang Wei
- Abstract summary: Capturing both local and global features of irregular point clouds is essential to 3D object detection (3OD)
This paper explores new modules to simultaneously learn local-global features of scene point clouds that serve 3OD positively.
We propose an effective 3OD network via simultaneous local-global feature learning (dubbed 3DLG-Detector)
- Score: 15.995277437128452
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Capturing both local and global features of irregular point clouds is
essential to 3D object detection (3OD). However, mainstream 3D detectors, e.g.,
VoteNet and its variants, either abandon considerable local features during
pooling operations or ignore many global features in the whole scene context.
This paper explores new modules to simultaneously learn local-global features
of scene point clouds that serve 3OD positively. To this end, we propose an
effective 3OD network via simultaneous local-global feature learning (dubbed
3DLG-Detector). 3DLG-Detector has two key contributions. First, it develops a
Dynamic Points Interaction (DPI) module that preserves effective local features
during pooling. Besides, DPI is detachable and can be incorporated into
existing 3OD networks to boost their performance. Second, it develops a Global
Context Aggregation module to aggregate multi-scale features from different
layers of the encoder to achieve scene context-awareness. Our method shows
improvements over thirteen competitors in terms of detection accuracy and
robustness on both the SUN RGB-D and ScanNet datasets. Source code will be
available upon publication.
Related papers
- PVAFN: Point-Voxel Attention Fusion Network with Multi-Pooling Enhancing for 3D Object Detection [59.355022416218624]
integration of point and voxel representations is becoming more common in LiDAR-based 3D object detection.
We propose a novel two-stage 3D object detector, called Point-Voxel Attention Fusion Network (PVAFN)
PVAFN uses a multi-pooling strategy to integrate both multi-scale and region-specific information effectively.
arXiv Detail & Related papers (2024-08-26T19:43:01Z) - Co-Fix3D: Enhancing 3D Object Detection with Collaborative Refinement [37.24731059950228]
3D object detection in driving scenarios faces the challenge of complex road environments.
We propose an advanced detection framework called Co-Fix3D.
Co-Fix3D integrates Local and Global Enhancement (LGE) modules to refine Bird's Eye View (BEV) features.
arXiv Detail & Related papers (2024-08-15T07:56:02Z) - Surface-biased Multi-Level Context 3D Object Detection [1.9723551683930771]
This work addresses the object detection task in 3D point clouds using a highly efficient, surface-biased, feature extraction method (wang2022rbgnet)
We propose a 3D object detector that extracts accurate feature representations of object candidates and leverages self-attention on point patches, object candidates, and on the global scene in 3D scene.
arXiv Detail & Related papers (2023-02-13T11:50:04Z) - CAGroup3D: Class-Aware Grouping for 3D Object Detection on Point Clouds [55.44204039410225]
We present a novel two-stage fully sparse convolutional 3D object detection framework, named CAGroup3D.
Our proposed method first generates some high-quality 3D proposals by leveraging the class-aware local group strategy on the object surface voxels.
To recover the features of missed voxels due to incorrect voxel-wise segmentation, we build a fully sparse convolutional RoI pooling module.
arXiv Detail & Related papers (2022-10-09T13:38:48Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - The Devil is in the Task: Exploiting Reciprocal Appearance-Localization
Features for Monocular 3D Object Detection [62.1185839286255]
Low-cost monocular 3D object detection plays a fundamental role in autonomous driving.
We introduce a Dynamic Feature Reflecting Network, named DFR-Net.
We rank 1st among all the monocular 3D object detectors in the KITTI test set.
arXiv Detail & Related papers (2021-12-28T07:31:18Z) - LATFormer: Locality-Aware Point-View Fusion Transformer for 3D Shape
Recognition [38.540048855119004]
We propose a novel Locality-Aware Point-View Fusion Transformer (LATFormer) for 3D shape retrieval and classification.
The core component of LATFormer is a module named Locality-Aware Fusion (LAF) which integrates the local features of correlated regions across the two modalities.
In our LATFormer, we utilize the LAF module to fuse the multi-scale features of the two modalities both bidirectionally and hierarchically to obtain more informative features.
arXiv Detail & Related papers (2021-09-03T03:23:27Z) - DH3D: Deep Hierarchical 3D Descriptors for Robust Large-Scale 6DoF
Relocalization [56.15308829924527]
We propose a Siamese network that jointly learns 3D local feature detection and description directly from raw 3D points.
For detecting 3D keypoints we predict the discriminativeness of the local descriptors in an unsupervised manner.
Experiments on various benchmarks demonstrate that our method achieves competitive results for both global point cloud retrieval and local point cloud registration.
arXiv Detail & Related papers (2020-07-17T20:21:22Z) - D3Feat: Joint Learning of Dense Detection and Description of 3D Local
Features [51.04841465193678]
We leverage a 3D fully convolutional network for 3D point clouds.
We propose a novel and practical learning mechanism that densely predicts both a detection score and a description feature for each 3D point.
Our method achieves state-of-the-art results in both indoor and outdoor scenarios.
arXiv Detail & Related papers (2020-03-06T12:51:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.