R-AGNO-RPN: A LIDAR-Camera Region Deep Network for Resolution-Agnostic
Detection
- URL: http://arxiv.org/abs/2012.05740v1
- Date: Thu, 10 Dec 2020 15:22:58 GMT
- Title: R-AGNO-RPN: A LIDAR-Camera Region Deep Network for Resolution-Agnostic
Detection
- Authors: Ruddy Th\'eodose, Dieumet Denis, Thierry Chateau, Vincent Fr\'emont,
Paul Checchin
- Abstract summary: R-AGNO-RPN, a region proposal network built on fusion of 3D point clouds and RGB images is proposed.
Our approach is designed to be also applied on low point cloud resolutions.
- Score: 3.4761212729163313
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current neural networks-based object detection approaches processing LiDAR
point clouds are generally trained from one kind of LiDAR sensors. However,
their performances decrease when they are tested with data coming from a
different LiDAR sensor than the one used for training, i.e., with a different
point cloud resolution. In this paper, R-AGNO-RPN, a region proposal network
built on fusion of 3D point clouds and RGB images is proposed for 3D object
detection regardless of point cloud resolution. As our approach is designed to
be also applied on low point cloud resolutions, the proposed method focuses on
object localization instead of estimating refined boxes on reduced data. The
resilience to low-resolution point cloud is obtained through image features
accurately mapped to Bird's Eye View and a specific data augmentation procedure
that improves the contribution of the RGB images. To show the proposed
network's ability to deal with different point clouds resolutions, experiments
are conducted on both data coming from the KITTI 3D Object Detection and the
nuScenes datasets. In addition, to assess its performances, our method is
compared to PointPillars, a well-known 3D detection network. Experimental
results show that even on point cloud data reduced by $80\%$ of its original
points, our method is still able to deliver relevant proposals localization.
Related papers
- Sparse-to-Dense LiDAR Point Generation by LiDAR-Camera Fusion for 3D Object Detection [9.076003184833557]
We propose the LiDAR-Camera Augmentation Network (LCANet), a novel framework that reconstructs LiDAR point cloud data by fusing 2D image features.
LCANet fuses data from LiDAR sensors by projecting image features into the 3D space, integrating semantic information into the point cloud data.
This fusion effectively compensates for LiDAR's weakness in detecting objects at long distances, which are often represented by sparse points.
arXiv Detail & Related papers (2024-09-23T13:03:31Z) - Uplifting Range-View-based 3D Semantic Segmentation in Real-Time with Multi-Sensor Fusion [18.431017678057348]
Range-View(RV)-based 3D point cloud segmentation is widely adopted due to its compact data form.
However, RV-based methods fall short in providing robust segmentation for the occluded points.
We propose a new LiDAR and Camera Range-view-based 3D point cloud semantic segmentation method (LaCRange)
In addition to being real-time, the proposed method achieves state-of-the-art results on nuScenes benchmark.
arXiv Detail & Related papers (2024-07-12T21:41:57Z) - AGO-Net: Association-Guided 3D Point Cloud Object Detection Network [86.10213302724085]
We propose a novel 3D detection framework that associates intact features for objects via domain adaptation.
We achieve new state-of-the-art performance on the KITTI 3D detection benchmark in both accuracy and speed.
arXiv Detail & Related papers (2022-08-24T16:54:38Z) - Open-Vocabulary 3D Detection via Image-level Class and Debiased
Cross-modal Contrastive Learning [62.18197846270103]
Current point-cloud detection methods have difficulty detecting the open-vocabulary objects in the real world.
We propose OV-3DETIC, an Open-Vocabulary 3D DETector using Image-level Class supervision.
arXiv Detail & Related papers (2022-07-05T12:13:52Z) - VPFNet: Improving 3D Object Detection with Virtual Point based LiDAR and
Stereo Data Fusion [62.24001258298076]
VPFNet is a new architecture that cleverly aligns and aggregates the point cloud and image data at the virtual' points.
Our VPFNet achieves 83.21% moderate 3D AP and 91.86% moderate BEV AP on the KITTI test set, ranking the 1st since May 21th, 2021.
arXiv Detail & Related papers (2021-11-29T08:51:20Z) - RoIFusion: 3D Object Detection from LiDAR and Vision [7.878027048763662]
We propose a novel fusion algorithm by projecting a set of 3D Region of Interests (RoIs) from the point clouds to the 2D RoIs of the corresponding the images.
Our approach achieves state-of-the-art performance on the KITTI 3D object detection challenging benchmark.
arXiv Detail & Related papers (2020-09-09T20:23:27Z) - Cross-Modality 3D Object Detection [63.29935886648709]
We present a novel two-stage multi-modal fusion network for 3D object detection.
The whole architecture facilitates two-stage fusion.
Our experiments on the KITTI dataset show that the proposed multi-stage fusion helps the network to learn better representations.
arXiv Detail & Related papers (2020-08-16T11:01:20Z) - Pseudo-LiDAR Point Cloud Interpolation Based on 3D Motion Representation
and Spatial Supervision [68.35777836993212]
We propose a Pseudo-LiDAR point cloud network to generate temporally and spatially high-quality point cloud sequences.
By exploiting the scene flow between point clouds, the proposed network is able to learn a more accurate representation of the 3D spatial motion relationship.
arXiv Detail & Related papers (2020-06-20T03:11:04Z) - GRNet: Gridding Residual Network for Dense Point Cloud Completion [54.43648460932248]
Estimating the complete 3D point cloud from an incomplete one is a key problem in many vision and robotics applications.
We propose a novel Gridding Residual Network (GRNet) for point cloud completion.
Experimental results indicate that the proposed GRNet performs favorably against state-of-the-art methods on the ShapeNet, Completion3D, and KITTI benchmarks.
arXiv Detail & Related papers (2020-06-06T02:46:39Z) - 3D Object Detection From LiDAR Data Using Distance Dependent Feature
Extraction [7.04185696830272]
This work proposes an improvement for 3D object detectors by taking into account the properties of LiDAR point clouds over distance.
Results show that training separate networks for close-range and long-range objects boosts performance for all KITTI benchmark difficulties.
arXiv Detail & Related papers (2020-03-02T13:16:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.