DurLAR: A High-fidelity 128-channel LiDAR Dataset with Panoramic Ambient and Reflectivity Imagery for Multi-modal Autonomous Driving Applications
- URL: http://arxiv.org/abs/2406.10068v1
- Date: Fri, 14 Jun 2024 14:24:05 GMT
- Title: DurLAR: A High-fidelity 128-channel LiDAR Dataset with Panoramic Ambient and Reflectivity Imagery for Multi-modal Autonomous Driving Applications
- Authors: Li Li, Khalid N. Ismail, Hubert P. H. Shum, Toby P. Breckon,
- Abstract summary: DurLAR is a high-fidelity 128-channel 3D LiDAR dataset with panoramic ambient (near infrared) and reflectivity imagery.
Our evaluation shows our joint use supervised and self-supervised loss terms, enabled via the superior ground truth resolution.
- Score: 21.066770408683265
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present DurLAR, a high-fidelity 128-channel 3D LiDAR dataset with panoramic ambient (near infrared) and reflectivity imagery, as well as a sample benchmark task using depth estimation for autonomous driving applications. Our driving platform is equipped with a high resolution 128 channel LiDAR, a 2MPix stereo camera, a lux meter and a GNSS/INS system. Ambient and reflectivity images are made available along with the LiDAR point clouds to facilitate multi-modal use of concurrent ambient and reflectivity scene information. Leveraging DurLAR, with a resolution exceeding that of prior benchmarks, we consider the task of monocular depth estimation and use this increased availability of higher resolution, yet sparse ground truth scene depth information to propose a novel joint supervised/self-supervised loss formulation. We compare performance over both our new DurLAR dataset, the established KITTI benchmark and the Cityscapes dataset. Our evaluation shows our joint use supervised and self-supervised loss terms, enabled via the superior ground truth resolution and availability within DurLAR improves the quantitative and qualitative performance of leading contemporary monocular depth estimation approaches (RMSE=3.639, Sq Rel=0.936).
Related papers
- LED: Light Enhanced Depth Estimation at Night [10.941842055797125]
We introduce Light Enhanced Depth (LED), a novel cost-effective approach that significantly improves depth estimation in low-light environments.
LED harnesses a pattern projected by high definition headlights available in modern vehicles.
We release the Nighttime Synthetic Drive dataset, which comprises 49,990 comprehensively annotated images.
arXiv Detail & Related papers (2024-09-12T13:23:24Z) - DINO-SD: Champion Solution for ICRA 2024 RoboDepth Challenge [54.71866583204417]
In this report, we introduce the DINO-SD, a novel surround-view depth estimation model.
Our DINO-SD does not need additional data and has strong robustness.
Our DINO-SD get the best performance in the track4 of ICRA 2024 RoboDepth Challenge.
arXiv Detail & Related papers (2024-05-27T12:21:31Z) - Better Monocular 3D Detectors with LiDAR from the Past [64.6759926054061]
Camera-based 3D detectors often suffer inferior performance compared to LiDAR-based counterparts due to inherent depth ambiguities in images.
In this work, we seek to improve monocular 3D detectors by leveraging unlabeled historical LiDAR data.
We show consistent and significant performance gain across multiple state-of-the-art models and datasets with a negligible additional latency of 9.66 ms and a small storage cost.
arXiv Detail & Related papers (2024-04-08T01:38:43Z) - RIDERS: Radar-Infrared Depth Estimation for Robust Sensing [22.10378524682712]
Adverse weather conditions pose significant challenges to accurate dense depth estimation.
We present a novel approach for robust metric depth estimation by fusing a millimeter-wave Radar and a monocular infrared thermal camera.
Our method achieves exceptional visual quality and accurate metric estimation by addressing the challenges of ambiguity and misalignment.
arXiv Detail & Related papers (2024-02-03T07:14:43Z) - Uncertainty Guided Depth Fusion for Spike Camera [49.41822923588663]
We propose a novel Uncertainty-Guided Depth Fusion (UGDF) framework to fuse predictions of monocular and stereo depth estimation networks for spike camera.
Our framework is motivated by the fact that stereo spike depth estimation achieves better results at close range.
In order to demonstrate the advantage of spike depth estimation over traditional camera depth estimation, we contribute a spike-depth dataset named CitySpike20K.
arXiv Detail & Related papers (2022-08-26T13:04:01Z) - Depth Estimation Matters Most: Improving Per-Object Depth Estimation for
Monocular 3D Detection and Tracking [47.59619420444781]
Approaches to monocular 3D perception including detection and tracking often yield inferior performance when compared to LiDAR-based techniques.
We propose a multi-level fusion method that combines different representations (RGB and pseudo-LiDAR) and temporal information across multiple frames for objects (tracklets) to enhance per-object depth estimation.
arXiv Detail & Related papers (2022-06-08T03:37:59Z) - LiDARCap: Long-range Marker-less 3D Human Motion Capture with LiDAR
Point Clouds [58.402752909624716]
Existing motion capture datasets are largely short-range and cannot yet fit the need of long-range applications.
We propose LiDARHuman26M, a new human motion capture dataset captured by LiDAR at a much longer range to overcome this limitation.
Our dataset also includes the ground truth human motions acquired by the IMU system and the synchronous RGB images.
arXiv Detail & Related papers (2022-03-28T12:52:45Z) - LiDARTouch: Monocular metric depth estimation with a few-beam LiDAR [40.98198236276633]
Vision-based depth estimation is a key feature in autonomous systems.
In such a monocular setup, dense depth is obtained with either additional input from one or several expensive LiDARs.
In this paper, we propose a new alternative of densely estimating metric depth by combining a monocular camera with a light-weight LiDAR.
arXiv Detail & Related papers (2021-09-08T12:06:31Z) - 3D Vehicle Detection Using Camera and Low-Resolution LiDAR [6.293059137498174]
We propose a novel framework for 3D object detection in Bird-Eye View (BEV) using a low-resolution LiDAR and a monocular camera.
Taking the low-resolution LiDAR point cloud and the monocular image as input, our depth completion network is able to produce dense point cloud.
For both easy and moderate cases, our detection results are comparable to those from 64-line high-resolution LiDAR.
arXiv Detail & Related papers (2021-05-04T21:08:20Z) - LEAD: LiDAR Extender for Autonomous Driving [48.233424487002445]
MEMS LiDAR emerges with irresistible trend due to its lower cost, more robust, and meeting the mass-production standards.
It suffers small field of view (FoV), slowing down the step of its population.
We propose LEAD, i.e., LiDAR Extender for Autonomous Driving, to extend the MEMS LiDAR by coupled image w.r.t both FoV and range.
arXiv Detail & Related papers (2021-02-16T07:35:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.