A comparison of uncertainty estimation approaches for DNN-based camera
localization
- URL: http://arxiv.org/abs/2211.01234v1
- Date: Wed, 2 Nov 2022 16:15:28 GMT
- Title: A comparison of uncertainty estimation approaches for DNN-based camera
localization
- Authors: Matteo Vaghi, Augusto Luis Ballardini, Simone Fontana, Domenico
Giorgio Sorrenti
- Abstract summary: This work compares the performances of three uncertainty estimation methods: Monte Carlo Dropout (MCD), Deep Ensemble (DE), and Deep Evidential Regression (DER)
We achieve accurate camera localization and a calibrated uncertainty, to the point that some method can be used for detecting localization failures.
- Score: 6.053739577423792
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Camera localization, i.e., camera pose regression, represents a very
important task in computer vision, since it has many practical applications,
such as autonomous driving. A reliable estimation of the uncertainties in
camera localization is also important, as it would allow to intercept
localization failures, which would be dangerous. Even though the literature
presents some uncertainty estimation methods, to the best of our knowledge
their effectiveness has not been thoroughly examined. This work compares the
performances of three consolidated epistemic uncertainty estimation methods:
Monte Carlo Dropout (MCD), Deep Ensemble (DE), and Deep Evidential Regression
(DER), in the specific context of camera localization. We exploited CMRNet, a
DNN approach for multi-modal image to LiDAR map registration, by modifying its
internal configuration to allow for an extensive experimental activity with the
three methods on the KITTI dataset. Particularly significant has been the
application of DER. We achieve accurate camera localization and a calibrated
uncertainty, to the point that some method can be used for detecting
localization failures.
Related papers
- SCIPaD: Incorporating Spatial Clues into Unsupervised Pose-Depth Joint Learning [17.99904937160487]
We introduce SCIPaD, a novel approach that incorporates spatial clues for unsupervised depth-pose joint learning.
SCIPaD achieves a reduction of 22.2% in average translation error and 34.8% in average angular error for camera pose estimation task on the KITTI Odometry dataset.
arXiv Detail & Related papers (2024-07-07T06:52:51Z) - Calib3D: Calibrating Model Preferences for Reliable 3D Scene Understanding [55.32861154245772]
Calib3D is a pioneering effort to benchmark and scrutinize the reliability of 3D scene understanding models.
We evaluate 28 state-of-the-art models across 10 diverse 3D datasets.
We introduce DeptS, a novel depth-aware scaling approach aimed at enhancing 3D model calibration.
arXiv Detail & Related papers (2024-03-25T17:59:59Z) - iComMa: Inverting 3D Gaussian Splatting for Camera Pose Estimation via Comparing and Matching [14.737266480464156]
We present a method named iComMa to address the 6D camera pose estimation problem in computer vision.
We propose an efficient method for accurate camera pose estimation by inverting 3D Gaussian Splatting (3DGS)
arXiv Detail & Related papers (2023-12-14T15:31:33Z) - P2O-Calib: Camera-LiDAR Calibration Using Point-Pair Spatial Occlusion
Relationship [1.6921147361216515]
We propose a novel target-less calibration approach based on the 2D-3D edge point extraction using the occlusion relationship in 3D space.
Our method achieves low error and high robustness that can contribute to the practical applications relying on high-quality Camera-LiDAR calibration.
arXiv Detail & Related papers (2023-11-04T14:32:55Z) - Uncertainty-Aware AB3DMOT by Variational 3D Object Detection [74.8441634948334]
Uncertainty estimation is an effective tool to provide statistically accurate predictions.
In this paper, we propose a Variational Neural Network-based TANet 3D object detector to generate 3D object detections with uncertainty.
arXiv Detail & Related papers (2023-02-12T14:30:03Z) - Improving Worst Case Visual Localization Coverage via Place-specific
Sub-selection in Multi-camera Systems [29.519262914510396]
6-DoF visual localization systems utilize principled approaches rooted in 3D geometry to perform accurate camera pose estimation of images to a map.
We demonstrate substantially improved worst-case localization performance compared to using off-the-shelf pipelines.
Our proposed approach is particularly applicable to the crowdsharing model of autonomous vehicle deployment.
arXiv Detail & Related papers (2022-06-28T10:59:39Z) - PDC-Net+: Enhanced Probabilistic Dense Correspondence Network [161.76275845530964]
Enhanced Probabilistic Dense Correspondence Network, PDC-Net+, capable of estimating accurate dense correspondences.
We develop an architecture and an enhanced training strategy tailored for robust and generalizable uncertainty prediction.
Our approach obtains state-of-the-art results on multiple challenging geometric matching and optical flow datasets.
arXiv Detail & Related papers (2021-09-28T17:56:41Z) - Uncertainty-Aware Camera Pose Estimation from Points and Lines [101.03675842534415]
Perspective-n-Point-and-Line (Pn$PL) aims at fast, accurate and robust camera localizations with respect to a 3D model from 2D-3D feature coordinates.
arXiv Detail & Related papers (2021-07-08T15:19:36Z) - Self-Supervised Multi-Frame Monocular Scene Flow [61.588808225321735]
We introduce a multi-frame monocular scene flow network based on self-supervised learning.
We observe state-of-the-art accuracy among monocular scene flow methods based on self-supervised learning.
arXiv Detail & Related papers (2021-05-05T17:49:55Z) - Delving into Localization Errors for Monocular 3D Object Detection [85.77319416168362]
Estimating 3D bounding boxes from monocular images is an essential component in autonomous driving.
In this work, we quantify the impact introduced by each sub-task and find the localization error' is the vital factor in restricting monocular 3D detection.
arXiv Detail & Related papers (2021-03-30T10:38:01Z) - Semi-Local 3D Lane Detection and Uncertainty Estimation [6.296104145657063]
Our method is based on a semi-local, BEV, tile representation that breaks down lanes into simple lane segments.
It combines learning a parametric model for the segments along with a deep feature embedding that is then used to cluster segment together into full lanes.
Our method is the first to output a learning based uncertainty estimation for the lane detection task.
arXiv Detail & Related papers (2020-03-11T12:35:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.