Calibrating LiDAR and Camera using Semantic Mutual information
- URL: http://arxiv.org/abs/2104.12023v1
- Date: Sat, 24 Apr 2021 21:04:33 GMT
- Title: Calibrating LiDAR and Camera using Semantic Mutual information
- Authors: Peng Jiang, Philip Osteen, Srikanth Saripalli
- Abstract summary: We propose an algorithm for automatic, targetless, extrinsic calibration of a LiDAR and camera system using semantic information.
We achieve this goal by maximizing mutual information (MI) of semantic information between sensors, leveraging a neural network to estimate semantic mutual information, and matrix exponential for calibration computation.
- Score: 8.40460868324361
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose an algorithm for automatic, targetless, extrinsic calibration of a
LiDAR and camera system using semantic information. We achieve this goal by
maximizing mutual information (MI) of semantic information between sensors,
leveraging a neural network to estimate semantic mutual information, and matrix
exponential for calibration computation. Using kernel-based sampling to sample
data from camera measurement based on LiDAR projected points, we formulate the
problem as a novel differentiable objective function which supports the use of
gradient-based optimization methods. We also introduce an initial calibration
method using 2D MI-based image registration. Finally, we demonstrate the
robustness of our method and quantitatively analyze the accuracy on a synthetic
dataset and also evaluate our algorithm qualitatively on KITTI360 and RELLIS-3D
benchmark datasets, showing improvement over recent comparable approaches.
Related papers
- YOCO: You Only Calibrate Once for Accurate Extrinsic Parameter in LiDAR-Camera Systems [0.5999777817331317]
In a multi-sensor fusion system composed of cameras and LiDAR, precise extrinsic calibration contributes to the system's long-term stability and accurate perception of the environment.
This paper proposes a novel fully automatic extrinsic calibration method for LiDAR-camera systems that circumvents the need for corresponding point registration.
arXiv Detail & Related papers (2024-07-25T13:44:49Z) - MM3DGS SLAM: Multi-modal 3D Gaussian Splatting for SLAM Using Vision, Depth, and Inertial Measurements [59.70107451308687]
We show for the first time that using 3D Gaussians for map representation with unposed camera images and inertial measurements can enable accurate SLAM.
Our method, MM3DGS, addresses the limitations of prior rendering by enabling faster scale awareness, and improved trajectory tracking.
We also release a multi-modal dataset, UT-MM, collected from a mobile robot equipped with a camera and an inertial measurement unit.
arXiv Detail & Related papers (2024-04-01T04:57:41Z) - Joint Spatial-Temporal Calibration for Camera and Global Pose Sensor [0.4143603294943439]
In robotics, motion capture systems have been widely used to measure the accuracy of localization algorithms.
These functionalities require having accurate and reliable spatial-temporal calibration parameters between the camera and the global pose sensor.
In this study, we provide two novel solutions to estimate these calibration parameters.
arXiv Detail & Related papers (2024-03-01T20:56:14Z) - GPS-Gaussian: Generalizable Pixel-wise 3D Gaussian Splatting for Real-time Human Novel View Synthesis [70.24111297192057]
We present a new approach, termed GPS-Gaussian, for synthesizing novel views of a character in a real-time manner.
The proposed method enables 2K-resolution rendering under a sparse-view camera setting.
arXiv Detail & Related papers (2023-12-04T18:59:55Z) - Multi-task Learning for Camera Calibration [3.274290296343038]
We present a unique method for predicting intrinsic (principal point offset and focal length) and extrinsic (baseline, pitch, and translation) properties from a pair of images.
By reconstructing the 3D points using a camera model neural network and then using the loss in reconstruction to obtain the camera specifications, this innovative camera projection loss (CPL) method allows us that the desired parameters should be estimated.
arXiv Detail & Related papers (2022-11-22T17:39:31Z) - Benchmarking the Robustness of LiDAR-Camera Fusion for 3D Object
Detection [58.81316192862618]
Two critical sensors for 3D perception in autonomous driving are the camera and the LiDAR.
fusing these two modalities can significantly boost the performance of 3D perception models.
We benchmark the state-of-the-art fusion methods for the first time.
arXiv Detail & Related papers (2022-05-30T09:35:37Z) - Data-Driven Interpolation for Super-Scarce X-Ray Computed Tomography [1.3535770763481902]
We train shallow neural networks to combine two neighbouring acquisitions into an estimated measurement at an intermediate angle.
This yields an enhanced sequence of measurements that can be reconstructed using standard methods.
Results are obtained for 2D and 3D imaging, on large biomedical datasets.
arXiv Detail & Related papers (2022-05-16T15:42:41Z) - LiDARCap: Long-range Marker-less 3D Human Motion Capture with LiDAR
Point Clouds [58.402752909624716]
Existing motion capture datasets are largely short-range and cannot yet fit the need of long-range applications.
We propose LiDARHuman26M, a new human motion capture dataset captured by LiDAR at a much longer range to overcome this limitation.
Our dataset also includes the ground truth human motions acquired by the IMU system and the synchronous RGB images.
arXiv Detail & Related papers (2022-03-28T12:52:45Z) - SemCal: Semantic LiDAR-Camera Calibration using Neural MutualInformation
Estimator [7.478076599395811]
SemCal is an automatic, targetless, extrinsic calibration algorithm for a LiDAR and camera system using semantic information.
We leverage a neural information estimator to estimate the mutual information (MI) of semantic information extracted from each sensor measurement.
arXiv Detail & Related papers (2021-09-21T15:51:24Z) - Automatic Extrinsic Calibration Method for LiDAR and Camera Sensor
Setups [68.8204255655161]
We present a method to calibrate the parameters of any pair of sensors involving LiDARs, monocular or stereo cameras.
The proposed approach can handle devices with very different resolutions and poses, as usually found in vehicle setups.
arXiv Detail & Related papers (2021-01-12T12:02:26Z) - Reinforced Axial Refinement Network for Monocular 3D Object Detection [160.34246529816085]
Monocular 3D object detection aims to extract the 3D position and properties of objects from a 2D input image.
Conventional approaches sample 3D bounding boxes from the space and infer the relationship between the target object and each of them, however, the probability of effective samples is relatively small in the 3D space.
We propose to start with an initial prediction and refine it gradually towards the ground truth, with only one 3d parameter changed in each step.
This requires designing a policy which gets a reward after several steps, and thus we adopt reinforcement learning to optimize it.
arXiv Detail & Related papers (2020-08-31T17:10:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.