Fast Uncertainty Quantification for Deep Object Pose Estimation
- URL: http://arxiv.org/abs/2011.07748v3
- Date: Fri, 26 Mar 2021 05:13:32 GMT
- Title: Fast Uncertainty Quantification for Deep Object Pose Estimation
- Authors: Guanya Shi, Yifeng Zhu, Jonathan Tremblay, Stan Birchfield, Fabio
Ramos, Animashree Anandkumar, Yuke Zhu
- Abstract summary: Deep learning-based object pose estimators are often unreliable and overconfident.
In this work, we propose a simple, efficient, and plug-and-play UQ method for 6-DoF object pose estimation.
- Score: 91.09217713805337
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep learning-based object pose estimators are often unreliable and
overconfident especially when the input image is outside the training domain,
for instance, with sim2real transfer. Efficient and robust uncertainty
quantification (UQ) in pose estimators is critically needed in many robotic
tasks. In this work, we propose a simple, efficient, and plug-and-play UQ
method for 6-DoF object pose estimation. We ensemble 2-3 pre-trained models
with different neural network architectures and/or training data sources, and
compute their average pairwise disagreement against one another to obtain the
uncertainty quantification. We propose four disagreement metrics, including a
learned metric, and show that the average distance (ADD) is the best
learning-free metric and it is only slightly worse than the learned metric,
which requires labeled target data. Our method has several advantages compared
to the prior art: 1) our method does not require any modification of the
training process or the model inputs; and 2) it needs only one forward pass for
each model. We evaluate the proposed UQ method on three tasks where our
uncertainty quantification yields much stronger correlations with pose
estimation errors than the baselines. Moreover, in a real robot grasping task,
our method increases the grasping success rate from 35% to 90%.
Related papers
- Distilling from Similar Tasks for Transfer Learning on a Budget [38.998980344852846]
Transfer learning is an effective solution for training with few labels, however often at the expense of a computationally costly fine-tuning of large base models.
We propose to mitigate this unpleasant trade-off between compute and accuracy via semi-supervised cross-domain distillation.
Our methods need no access to source data, and merely need features and pseudo-labels of the source models.
arXiv Detail & Related papers (2023-04-24T17:59:01Z) - Post-hoc Uncertainty Learning using a Dirichlet Meta-Model [28.522673618527417]
We propose a novel Bayesian meta-model to augment pre-trained models with better uncertainty quantification abilities.
Our proposed method requires no additional training data and is flexible enough to quantify different uncertainties.
We demonstrate our proposed meta-model approach's flexibility and superior empirical performance on these applications.
arXiv Detail & Related papers (2022-12-14T17:34:11Z) - What Stops Learning-based 3D Registration from Working in the Real
World? [53.68326201131434]
This work identifies the sources of 3D point cloud registration failures, analyze the reasons behind them, and propose solutions.
Ultimately, this translates to a best-practice 3D registration network (BPNet), constituting the first learning-based method able to handle previously-unseen objects in real-world data.
Our model generalizes to real data without any fine-tuning, reaching an accuracy of up to 67% on point clouds of unseen objects obtained with a commercial sensor.
arXiv Detail & Related papers (2021-11-19T19:24:27Z) - Geometry Uncertainty Projection Network for Monocular 3D Object
Detection [138.24798140338095]
We propose a Geometry Uncertainty Projection Network (GUP Net) to tackle the error amplification problem at both inference and training stages.
Specifically, a GUP module is proposed to obtains the geometry-guided uncertainty of the inferred depth.
At the training stage, we propose a Hierarchical Task Learning strategy to reduce the instability caused by error amplification.
arXiv Detail & Related papers (2021-07-29T06:59:07Z) - SIMPLE: SIngle-network with Mimicking and Point Learning for Bottom-up
Human Pose Estimation [81.03485688525133]
We propose a novel multi-person pose estimation framework, SIngle-network with Mimicking and Point Learning for Bottom-up Human Pose Estimation (SIMPLE)
Specifically, in the training process, we enable SIMPLE to mimic the pose knowledge from the high-performance top-down pipeline.
Besides, SIMPLE formulates human detection and pose estimation as a unified point learning framework to complement each other in single-network.
arXiv Detail & Related papers (2021-04-06T13:12:51Z) - The Right Tool for the Job: Matching Model and Instance Complexities [62.95183777679024]
As NLP models become larger, executing a trained model requires significant computational resources incurring monetary and environmental costs.
We propose a modification to contextual representation fine-tuning which, during inference, allows for an early (and fast) "exit"
We test our proposed modification on five different datasets in two tasks: three text classification datasets and two natural language inference benchmarks.
arXiv Detail & Related papers (2020-04-16T04:28:08Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.