Sparse LiDAR and Stereo Fusion (SLS-Fusion) for Depth Estimationand 3D
Object Detection
- URL: http://arxiv.org/abs/2103.03977v1
- Date: Fri, 5 Mar 2021 23:10:09 GMT
- Title: Sparse LiDAR and Stereo Fusion (SLS-Fusion) for Depth Estimationand 3D
Object Detection
- Authors: Nguyen Anh Minh Mai, Pierre Duthon, Louahdi Khoudour, Alain Crouzil,
Sergio A. Velastin
- Abstract summary: SLS-Fusion is a new approach to fuse data from 4-beam LiDAR and a stereo camera via a neural network for depth estimation.
Since 4-beam LiDAR is cheaper than the well-known 64-beam LiDAR, this approach is also classified as a low-cost sensors-based method.
- Score: 3.5488685789514736
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The ability to accurately detect and localize objects is recognized as being
the most important for the perception of self-driving cars. From 2D to 3D
object detection, the most difficult is to determine the distance from the
ego-vehicle to objects. Expensive technology like LiDAR can provide a precise
and accurate depth information, so most studies have tended to focus on this
sensor showing a performance gap between LiDAR-based methods and camera-based
methods. Although many authors have investigated how to fuse LiDAR with RGB
cameras, as far as we know there are no studies to fuse LiDAR and stereo in a
deep neural network for the 3D object detection task. This paper presents
SLS-Fusion, a new approach to fuse data from 4-beam LiDAR and a stereo camera
via a neural network for depth estimation to achieve better dense depth maps
and thereby improves 3D object detection performance. Since 4-beam LiDAR is
cheaper than the well-known 64-beam LiDAR, this approach is also classified as
a low-cost sensors-based method. Through evaluation on the KITTI benchmark, it
is shown that the proposed method significantly improves depth estimation
performance compared to a baseline method. Also, when applying it to 3D object
detection, a new state of the art on low-cost sensor based method is achieved.
Related papers
- SimpleBEV: Improved LiDAR-Camera Fusion Architecture for 3D Object Detection [15.551625571158056]
We propose a LiDAR-camera fusion framework, named SimpleBEV, for accurate 3D object detection.
Our method achieves 77.6% NDS accuracy on the nuScenes dataset, showcasing superior performance in the 3D object detection track.
arXiv Detail & Related papers (2024-11-08T02:51:39Z) - Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - MonoDistill: Learning Spatial Features for Monocular 3D Object Detection [80.74622486604886]
We propose a simple and effective scheme to introduce the spatial information from LiDAR signals to the monocular 3D detectors.
We use the resulting data to train a 3D detector with the same architecture as the baseline model.
Experimental results show that the proposed method can significantly boost the performance of the baseline model.
arXiv Detail & Related papers (2022-01-26T09:21:41Z) - SIDE: Center-based Stereo 3D Detector with Structure-aware Instance
Depth Estimation [11.169586369931803]
We propose a stereo-image based anchor-free 3D detection method, called structure-aware stereo 3D detector (termed as SIDE)
We explore the instance-level depth information via constructing the cost volume from RoIs of each object.
Our method achieves the state-of-the-art performance compared to existing methods without depth map supervision.
arXiv Detail & Related papers (2021-08-22T08:09:38Z) - LIGA-Stereo: Learning LiDAR Geometry Aware Representations for
Stereo-based 3D Detector [80.7563981951707]
We propose LIGA-Stereo to learn stereo-based 3D detectors under the guidance of high-level geometry-aware representations of LiDAR-based detection models.
Compared with the state-of-the-art stereo detector, our method has improved the 3D detection performance of cars, pedestrians, cyclists by 10.44%, 5.69%, 5.97% mAP respectively.
arXiv Detail & Related papers (2021-08-18T17:24:40Z) - CFTrack: Center-based Radar and Camera Fusion for 3D Multi-Object
Tracking [9.62721286522053]
We propose an end-to-end network for joint object detection and tracking based on radar and camera sensor fusion.
Our proposed method uses a center-based radar-camera fusion algorithm for object detection and utilizes a greedy algorithm for object association.
We evaluate our method on the challenging nuScenes dataset, where it achieves 20.0 AMOTA and outperforms all vision-based 3D tracking methods in the benchmark.
arXiv Detail & Related papers (2021-07-11T23:56:53Z) - 3D Vehicle Detection Using Camera and Low-Resolution LiDAR [6.293059137498174]
We propose a novel framework for 3D object detection in Bird-Eye View (BEV) using a low-resolution LiDAR and a monocular camera.
Taking the low-resolution LiDAR point cloud and the monocular image as input, our depth completion network is able to produce dense point cloud.
For both easy and moderate cases, our detection results are comparable to those from 64-line high-resolution LiDAR.
arXiv Detail & Related papers (2021-05-04T21:08:20Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - End-to-End Pseudo-LiDAR for Image-Based 3D Object Detection [62.34374949726333]
Pseudo-LiDAR (PL) has led to a drastic reduction in the accuracy gap between methods based on LiDAR sensors and those based on cheap stereo cameras.
PL combines state-of-the-art deep neural networks for 3D depth estimation with those for 3D object detection by converting 2D depth map outputs to 3D point cloud inputs.
We introduce a new framework based on differentiable Change of Representation (CoR) modules that allow the entire PL pipeline to be trained end-to-end.
arXiv Detail & Related papers (2020-04-07T02:18:38Z) - DSGN: Deep Stereo Geometry Network for 3D Object Detection [79.16397166985706]
There is a large performance gap between image-based and LiDAR-based 3D object detectors.
Our method, called Deep Stereo Geometry Network (DSGN), significantly reduces this gap.
For the first time, we provide a simple and effective one-stage stereo-based 3D detection pipeline.
arXiv Detail & Related papers (2020-01-10T11:44:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.