CalibDNN: Multimodal Sensor Calibration for Perception Using Deep Neural
Networks
- URL: http://arxiv.org/abs/2103.14793v1
- Date: Sat, 27 Mar 2021 02:43:37 GMT
- Title: CalibDNN: Multimodal Sensor Calibration for Perception Using Deep Neural
Networks
- Authors: Ganning Zhao, Jiesi Hu, Suya You and C.-C. Jay Kuo
- Abstract summary: We propose a novel deep learning-driven technique (CalibDNN) for accurate calibration among multimodal sensor, specifically LiDAR-Camera pairs.
The entire processing is fully automatic with a single model and single iteration.
Results comparison among different methods and extensive experiments on different datasets demonstrates the state-of-the-art performance.
- Score: 27.877734292570967
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Current perception systems often carry multimodal imagers and sensors such as
2D cameras and 3D LiDAR sensors. To fuse and utilize the data for downstream
perception tasks, robust and accurate calibration of the multimodal sensor data
is essential. We propose a novel deep learning-driven technique (CalibDNN) for
accurate calibration among multimodal sensor, specifically LiDAR-Camera pairs.
The key innovation of the proposed work is that it does not require any
specific calibration targets or hardware assistants, and the entire processing
is fully automatic with a single model and single iteration. Results comparison
among different methods and extensive experiments on different datasets
demonstrates the state-of-the-art performance.
Related papers
- MSSIDD: A Benchmark for Multi-Sensor Denoising [55.41612200877861]
We introduce a new benchmark, the Multi-Sensor SIDD dataset, which is the first raw-domain dataset designed to evaluate the sensor transferability of denoising models.
We propose a sensor consistency training framework that enables denoising models to learn the sensor-invariant features.
arXiv Detail & Related papers (2024-11-18T13:32:59Z) - Multimodal Object Detection using Depth and Image Data for Manufacturing Parts [1.0819408603463427]
This work proposes a multi-sensor system combining an red-green-blue (RGB) camera and a 3D point cloud sensor.
A novel multimodal object detection method is developed to process both RGB and depth data.
The results show that the multimodal model significantly outperforms the depth-only and RGB-only baselines on established object detection metrics.
arXiv Detail & Related papers (2024-11-13T22:43:15Z) - UniCal: Unified Neural Sensor Calibration [32.7372115947273]
Self-driving vehicles (SDVs) require accurate calibration of LiDARs and cameras to fuse sensor data accurately for autonomy.
Traditional calibration methods leverage fiducials captured in a controlled and structured scene and compute correspondences to optimize over.
We propose UniCal, a unified framework for effortlessly calibrating SDVs equipped with multiple LiDARs and cameras.
arXiv Detail & Related papers (2024-09-27T17:56:04Z) - Sparse Points to Dense Clouds: Enhancing 3D Detection with Limited LiDAR Data [68.18735997052265]
We propose a balanced approach that combines the advantages of monocular and point cloud-based 3D detection.
Our method requires only a small number of 3D points, that can be obtained from a low-cost, low-resolution sensor.
The accuracy of 3D detection improves by 20% compared to the state-of-the-art monocular detection methods.
arXiv Detail & Related papers (2024-04-10T03:54:53Z) - SOAC: Spatio-Temporal Overlap-Aware Multi-Sensor Calibration using Neural Radiance Fields [10.958143040692141]
In rapidly-evolving domains such as autonomous driving, the use of multiple sensors with different modalities is crucial to ensure operational precision and stability.
To correctly exploit the provided information by each sensor in a single common frame, it is essential for these sensors to be accurately calibrated.
We leverage the ability of Neural Radiance Fields to represent different modalities in a common representation.
arXiv Detail & Related papers (2023-11-27T13:25:47Z) - Multi-Modal Multi-Task (3MT) Road Segmentation [0.8287206589886879]
We focus on using raw sensor inputs instead of, as it is typically done in many SOTA works, leveraging architectures that require high pre-processing costs.
This study presents a cost-effective and highly accurate solution for road segmentation by integrating data from multiple sensors within a multi-task learning architecture.
arXiv Detail & Related papers (2023-08-23T08:15:15Z) - Multi-Modal 3D Object Detection by Box Matching [109.43430123791684]
We propose a novel Fusion network by Box Matching (FBMNet) for multi-modal 3D detection.
With the learned assignments between 3D and 2D object proposals, the fusion for detection can be effectively performed by combing their ROI features.
arXiv Detail & Related papers (2023-05-12T18:08:51Z) - Learning Online Multi-Sensor Depth Fusion [100.84519175539378]
SenFuNet is a depth fusion approach that learns sensor-specific noise and outlier statistics.
We conduct experiments with various sensor combinations on the real-world CoRBS and Scene3D datasets.
arXiv Detail & Related papers (2022-04-07T10:45:32Z) - Automatic Extrinsic Calibration Method for LiDAR and Camera Sensor
Setups [68.8204255655161]
We present a method to calibrate the parameters of any pair of sensors involving LiDARs, monocular or stereo cameras.
The proposed approach can handle devices with very different resolutions and poses, as usually found in vehicle setups.
arXiv Detail & Related papers (2021-01-12T12:02:26Z) - Learning Camera Miscalibration Detection [83.38916296044394]
This paper focuses on a data-driven approach to learn the detection of miscalibration in vision sensors, specifically RGB cameras.
Our contributions include a proposed miscalibration metric for RGB cameras and a novel semi-synthetic dataset generation pipeline based on this metric.
By training a deep convolutional neural network, we demonstrate the effectiveness of our pipeline to identify whether a recalibration of the camera's intrinsic parameters is required or not.
arXiv Detail & Related papers (2020-05-24T10:32:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.