MLOD: Awareness of Extrinsic Perturbation in Multi-LiDAR 3D Object
Detection for Autonomous Driving
- URL: http://arxiv.org/abs/2010.11702v1
- Date: Tue, 29 Sep 2020 06:11:22 GMT
- Title: MLOD: Awareness of Extrinsic Perturbation in Multi-LiDAR 3D Object
Detection for Autonomous Driving
- Authors: Jianhao Jiao, Peng Yun, Lei Tai, Ming Liu
- Abstract summary: Extrinsic perturbation always exists in multiple sensors.
We propose a multi-LiDAR 3D object detector called MLOD.
We conduct extensive experiments on a real-world dataset.
- Score: 10.855519369371853
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Extrinsic perturbation always exists in multiple sensors. In this paper, we
focus on the extrinsic uncertainty in multi-LiDAR systems for 3D object
detection. We first analyze the influence of extrinsic perturbation on
geometric tasks with two basic examples. To minimize the detrimental effect of
extrinsic perturbation, we propagate an uncertainty prior on each point of
input point clouds, and use this information to boost an approach for 3D
geometric tasks. Then we extend our findings to propose a multi-LiDAR 3D object
detector called MLOD. MLOD is a two-stage network where the multi-LiDAR
information is fused through various schemes in stage one, and the extrinsic
perturbation is handled in stage two. We conduct extensive experiments on a
real-world dataset, and demonstrate both the accuracy and robustness
improvement of MLOD. The code, data and supplementary materials are available
at: https://ram-lab.com/file/site/mlod
Related papers
- STONE: A Submodular Optimization Framework for Active 3D Object Detection [20.54906045954377]
Key requirement for training an accurate 3D object detector is the availability of a large amount of LiDAR-based point cloud data.
This paper proposes a unified active 3D object detection framework, for greatly reducing the labeling cost of training 3D object detectors.
arXiv Detail & Related papers (2024-10-04T20:45:33Z) - LLMI3D: Empowering LLM with 3D Perception from a Single 2D Image [72.14973729674995]
Current 3D perception methods, particularly small models, struggle with processing logical reasoning, question-answering, and handling open scenario categories.
We propose solutions: Spatial-Enhanced Local Feature Mining for better spatial feature extraction, 3D Query Token-Derived Info Decoding for precise geometric regression, and Geometry Projection-Based 3D Reasoning for handling camera focal length variations.
arXiv Detail & Related papers (2024-08-14T10:00:16Z) - Towards Unified 3D Object Detection via Algorithm and Data Unification [70.27631528933482]
We build the first unified multi-modal 3D object detection benchmark MM- Omni3D and extend the aforementioned monocular detector to its multi-modal version.
We name the designed monocular and multi-modal detectors as UniMODE and MM-UniMODE, respectively.
arXiv Detail & Related papers (2024-02-28T18:59:31Z) - MSF3DDETR: Multi-Sensor Fusion 3D Detection Transformer for Autonomous
Driving [0.0]
We propose MSF3DDETR: Multi-Sensor Fusion 3D Detection Transformer architecture to fuse image and LiDAR features to improve the detection accuracy.
Our end-to-end single-stage, anchor-free and NMS-free network takes in multi-view images and LiDAR point clouds and predicts 3D bounding boxes.
MSF3DDETR network is trained end-to-end on the nuScenes dataset using Hungarian algorithm based bipartite matching and set-to-set loss inspired by DETR.
arXiv Detail & Related papers (2022-10-27T10:55:15Z) - M$^2$-3DLaneNet: Exploring Multi-Modal 3D Lane Detection [30.250833348463633]
M$2$-3DLaneNet lifts 2D features into 3D space by incorporating geometry information from LiDAR data through depth completion.
Experiments on the large-scale OpenLane dataset demonstrate the effectiveness of M$2$-3DLaneNet, regardless of the range.
arXiv Detail & Related papers (2022-09-13T13:45:18Z) - Multi-Echo LiDAR for 3D Object Detection [29.690900492033578]
A single laser pulse can be partially reflected by multiple objects along its path, resulting in multiple measurements called echoes.
LiDAR can also measure surface reflectance (intensity of laser pulse return), as well as ambient light of the scene.
We present a 3D object detection model which leverages the full spectrum of measurement signals provided by LiDAR.
arXiv Detail & Related papers (2021-07-23T21:43:09Z) - Delving into Localization Errors for Monocular 3D Object Detection [85.77319416168362]
Estimating 3D bounding boxes from monocular images is an essential component in autonomous driving.
In this work, we quantify the impact introduced by each sub-task and find the localization error' is the vital factor in restricting monocular 3D detection.
arXiv Detail & Related papers (2021-03-30T10:38:01Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - SelfVoxeLO: Self-supervised LiDAR Odometry with Voxel-based Deep Neural
Networks [81.64530401885476]
We propose a self-supervised LiDAR odometry method, dubbed SelfVoxeLO, to tackle these two difficulties.
Specifically, we propose a 3D convolution network to process the raw LiDAR data directly, which extracts features that better encode the 3D geometric patterns.
We evaluate our method's performances on two large-scale datasets, i.e., KITTI and Apollo-SouthBay.
arXiv Detail & Related papers (2020-10-19T09:23:39Z) - DSGN: Deep Stereo Geometry Network for 3D Object Detection [79.16397166985706]
There is a large performance gap between image-based and LiDAR-based 3D object detectors.
Our method, called Deep Stereo Geometry Network (DSGN), significantly reduces this gap.
For the first time, we provide a simple and effective one-stage stereo-based 3D detection pipeline.
arXiv Detail & Related papers (2020-01-10T11:44:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.