Uncertainty Quantification with Deep Ensembles for 6D Object Pose Estimation
- URL: http://arxiv.org/abs/2403.07741v2
- Date: Thu, 2 May 2024 09:00:21 GMT
- Title: Uncertainty Quantification with Deep Ensembles for 6D Object Pose Estimation
- Authors: Kira Wursthorn, Markus Hillemann, Markus Ulrich,
- Abstract summary: We propose a method to quantify the uncertainty of multi-stage 6D object pose estimation approaches with deep ensembles.
For the implementation, we choose SurfEmb as representative, since it is one of the top-performing 6D object pose estimation approaches.
- Score: 4.2603120588176635
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The estimation of 6D object poses is a fundamental task in many computer vision applications. Particularly, in high risk scenarios such as human-robot interaction, industrial inspection, and automation, reliable pose estimates are crucial. In the last years, increasingly accurate and robust deep-learning-based approaches for 6D object pose estimation have been proposed. Many top-performing methods are not end-to-end trainable but consist of multiple stages. In the context of deep uncertainty quantification, deep ensembles are considered as state of the art since they have been proven to produce well-calibrated and robust uncertainty estimates. However, deep ensembles can only be applied to methods that can be trained end-to-end. In this work, we propose a method to quantify the uncertainty of multi-stage 6D object pose estimation approaches with deep ensembles. For the implementation, we choose SurfEmb as representative, since it is one of the top-performing 6D object pose estimation approaches in the BOP Challenge 2022. We apply established metrics and concepts for deep uncertainty quantification to evaluate the results. Furthermore, we propose a novel uncertainty calibration score for regression tasks to quantify the quality of the estimated uncertainty.
Related papers
- Consensus-Driven Uncertainty for Robotic Grasping based on RGB Perception [4.719664724709857]
A grasping agent that both estimates the 6-DoF pose of a target object and predicts the uncertainty of its own estimate could avoid task failure by choosing not to act under high uncertainty.<n>We propose a method for training lightweight, deep networks to predict whether a grasp guided by an image-based pose estimate will succeed before that grasp is attempted.
arXiv Detail & Related papers (2025-06-24T22:53:54Z) - A Critical Synthesis of Uncertainty Quantification and Foundation Models in Monocular Depth Estimation [13.062551984263031]
Metric depth estimation, which involves predicting absolute distances, poses particular challenges.
We fuse five different uncertainty quantification methods with the current state-of-the-art DepthAnythingV2 foundation model.
Our findings identify fine-tuning with the Gaussian Negative Log-Likelihood Loss (GNLL) as a particularly promising approach.
arXiv Detail & Related papers (2025-01-14T15:13:00Z) - MaskVal: Simple but Effective Uncertainty Quantification for 6D Pose Estimation [0.0]
We investigate a simple but effective uncertainty quantification, that we call MaskVal, for 6D pose estimation.
Despite its simplicity, MaskVal significantly outperforms a state-of-the-art ensemble method on both a dataset and a robotic setup.
We show that by using MaskVal, the performance of a state-of-the-art 6D pose estimator is significantly improved towards a safe and reliable operation.
arXiv Detail & Related papers (2024-09-05T14:17:01Z) - Modeling the Uncertainty with Maximum Discrepant Students for
Semi-supervised 2D Pose Estimation [57.17120203327993]
We propose a framework to estimate the quality of pseudo-labels in semi-supervised pose estimation tasks.
Our method improves the performance of semi-supervised pose estimation on three datasets.
arXiv Detail & Related papers (2023-11-03T08:11:06Z) - Learning to Estimate 6DoF Pose from Limited Data: A Few-Shot,
Generalizable Approach using RGB Images [60.0898989456276]
We present a new framework named Cas6D for few-shot 6DoF pose estimation that is generalizable and uses only RGB images.
To address the false positives of target object detection in the extreme few-shot setting, our framework utilizes a self-supervised pre-trained ViT to learn robust feature representations.
Experimental results on the LINEMOD and GenMOP datasets demonstrate that Cas6D outperforms state-of-the-art methods by 9.2% and 3.8% accuracy (Proj-5) under the 32-shot setting.
arXiv Detail & Related papers (2023-06-13T07:45:42Z) - Rigidity-Aware Detection for 6D Object Pose Estimation [60.88857851869196]
Most recent 6D object pose estimation methods first use object detection to obtain 2D bounding boxes before actually regressing the pose.
We propose a rigidity-aware detection method exploiting the fact that, in 6D pose estimation, the target objects are rigid.
Key to the success of our approach is a visibility map, which we propose to build using a minimum barrier distance between every pixel in the bounding box and the box boundary.
arXiv Detail & Related papers (2023-03-22T09:02:54Z) - Ki-Pode: Keypoint-based Implicit Pose Distribution Estimation of Rigid
Objects [1.209625228546081]
We propose a novel pose distribution estimation method.
An implicit formulation of the probability distribution over object pose is derived from an intermediary representation of an object as a set of keypoints.
The method has been evaluated on the task of rotation distribution estimation on the YCB-V and T-LESS datasets.
arXiv Detail & Related papers (2022-09-20T11:59:05Z) - Instance-specific 6-DoF Object Pose Estimation from Minimal Annotations [6.24717069374781]
We present a method to rapidly train and deploy a pipeline for estimating the continuous 6-DoF pose of an object from a single RGB image.
The key idea is to leverage known camera poses and rigid body geometry to partially automate the generation of a large labeled dataset.
The dataset, along with sufficient domain randomization, is then used to supervise the training of deep neural networks for predicting semantic keypoints.
arXiv Detail & Related papers (2022-07-27T03:00:28Z) - BayesCap: Bayesian Identity Cap for Calibrated Uncertainty in Frozen
Neural Networks [50.15201777970128]
We propose BayesCap that learns a Bayesian identity mapping for the frozen model, allowing uncertainty estimation.
BayesCap is a memory-efficient method that can be trained on a small fraction of the original dataset.
We show the efficacy of our method on a wide variety of tasks with a diverse set of architectures.
arXiv Detail & Related papers (2022-07-14T12:50:09Z) - Unseen Object 6D Pose Estimation: A Benchmark and Baselines [62.8809734237213]
We propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
We collect a dataset with both real and synthetic images and up to 48 unseen objects in the test set.
By training an end-to-end 3D correspondences network, our method finds corresponding points between an unseen object and a partial view RGBD image accurately and efficiently.
arXiv Detail & Related papers (2022-06-23T16:29:53Z) - An Uncertainty Estimation Framework for Probabilistic Object Detection [5.83620245905973]
We introduce a new technique that combines two popular methods to estimate uncertainty in object detection.
Our framework employs deep ensembles and Monte Carlo dropout for approximating predictive uncertainty.
arXiv Detail & Related papers (2021-06-28T22:29:59Z) - Spatial Attention Improves Iterative 6D Object Pose Estimation [52.365075652976735]
We propose a new method for 6D pose estimation refinement from RGB images.
Our main insight is that after the initial pose estimate, it is important to pay attention to distinct spatial features of the object.
We experimentally show that this approach learns to attend to salient spatial features and learns to ignore occluded parts of the object, leading to better pose estimation across datasets.
arXiv Detail & Related papers (2021-01-05T17:18:52Z) - REDE: End-to-end Object 6D Pose Robust Estimation Using Differentiable
Outliers Elimination [15.736699709454857]
We propose REDE, a novel end-to-end object pose estimator using RGB-D data.
We also propose a differentiable outliers elimination method that regresses the candidate result and the confidence simultaneously.
The experimental results on three benchmark datasets show that REDE slightly outperforms the state-of-the-art approaches.
arXiv Detail & Related papers (2020-10-24T06:45:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.