Consensus-Driven Uncertainty for Robotic Grasping based on RGB Perception
- URL: http://arxiv.org/abs/2506.20045v2
- Date: Thu, 26 Jun 2025 15:12:14 GMT
- Title: Consensus-Driven Uncertainty for Robotic Grasping based on RGB Perception
- Authors: Eric C. Joyce, Qianwen Zhao, Nathaniel Burgdorfer, Long Wang, Philippos Mordohai,
- Abstract summary: A grasping agent that both estimates the 6-DoF pose of a target object and predicts the uncertainty of its own estimate could avoid task failure by choosing not to act under high uncertainty.<n>We propose a method for training lightweight, deep networks to predict whether a grasp guided by an image-based pose estimate will succeed before that grasp is attempted.
- Score: 4.719664724709857
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep object pose estimators are notoriously overconfident. A grasping agent that both estimates the 6-DoF pose of a target object and predicts the uncertainty of its own estimate could avoid task failure by choosing not to act under high uncertainty. Even though object pose estimation improves and uncertainty quantification research continues to make strides, few studies have connected them to the downstream task of robotic grasping. We propose a method for training lightweight, deep networks to predict whether a grasp guided by an image-based pose estimate will succeed before that grasp is attempted. We generate training data for our networks via object pose estimation on real images and simulated grasping. We also find that, despite high object variability in grasping trials, networks benefit from training on all objects jointly, suggesting that a diverse variety of objects can nevertheless contribute to the same goal.
Related papers
- Post-hoc Probabilistic Vision-Language Models [51.12284891724463]
Vision-language models (VLMs) have found remarkable success in classification, retrieval, and generative tasks.<n>We propose post-hoc uncertainty estimation in VLMs that does not require additional training.<n>Our results show promise for safety-critical applications of large-scale models.
arXiv Detail & Related papers (2024-12-08T18:16:13Z) - Uncertainty Quantification with Deep Ensembles for 6D Object Pose Estimation [4.2603120588176635]
We propose a method to quantify the uncertainty of multi-stage 6D object pose estimation approaches with deep ensembles.
For the implementation, we choose SurfEmb as representative, since it is one of the top-performing 6D object pose estimation approaches.
arXiv Detail & Related papers (2024-03-12T15:19:25Z) - Ambiguity-Aware Multi-Object Pose Optimization for Visually-Assisted
Robot Manipulation [17.440729138126162]
We present an ambiguity-aware 6D object pose estimation network, PrimA6D++, as a generic uncertainty prediction method.
The proposed method shows a significant performance improvement in T-LESS and YCB-Video datasets.
We further demonstrate real-time scene recognition capability for visually-assisted robot manipulation.
arXiv Detail & Related papers (2022-11-02T08:57:20Z) - Ki-Pode: Keypoint-based Implicit Pose Distribution Estimation of Rigid
Objects [1.209625228546081]
We propose a novel pose distribution estimation method.
An implicit formulation of the probability distribution over object pose is derived from an intermediary representation of an object as a set of keypoints.
The method has been evaluated on the task of rotation distribution estimation on the YCB-V and T-LESS datasets.
arXiv Detail & Related papers (2022-09-20T11:59:05Z) - Object Detection in Aerial Images with Uncertainty-Aware Graph Network [61.02591506040606]
We propose a novel uncertainty-aware object detection framework with a structured-graph, where nodes and edges are denoted by objects.
We refer to our model as Uncertainty-Aware Graph network for object DETection (UAGDet)
arXiv Detail & Related papers (2022-08-23T07:29:03Z) - Gradient-based Uncertainty for Monocular Depth Estimation [5.7575052885308455]
In monocular depth estimation, disturbances in the image context, like moving objects or reflecting materials, can easily lead to erroneous predictions.
We propose a post hoc uncertainty estimation approach for an already trained and thus fixed depth estimation model.
Our approach achieves state-of-the-art uncertainty estimation results on the KITTI and NYU Depth V2 benchmarks without the need to retrain the neural network.
arXiv Detail & Related papers (2022-08-03T12:21:02Z) - Instance-specific 6-DoF Object Pose Estimation from Minimal Annotations [6.24717069374781]
We present a method to rapidly train and deploy a pipeline for estimating the continuous 6-DoF pose of an object from a single RGB image.
The key idea is to leverage known camera poses and rigid body geometry to partially automate the generation of a large labeled dataset.
The dataset, along with sufficient domain randomization, is then used to supervise the training of deep neural networks for predicting semantic keypoints.
arXiv Detail & Related papers (2022-07-27T03:00:28Z) - CertainNet: Sampling-free Uncertainty Estimation for Object Detection [65.28989536741658]
Estimating the uncertainty of a neural network plays a fundamental role in safety-critical settings.
In this work, we propose a novel sampling-free uncertainty estimation method for object detection.
We call it CertainNet, and it is the first to provide separate uncertainties for each output signal: objectness, class, location and size.
arXiv Detail & Related papers (2021-10-04T17:59:31Z) - Learning Dynamics via Graph Neural Networks for Human Pose Estimation
and Tracking [98.91894395941766]
We propose a novel online approach to learning the pose dynamics, which are independent of pose detections in current fame.
Specifically, we derive this prediction of dynamics through a graph neural network(GNN) that explicitly accounts for both spatial-temporal and visual information.
Experiments on PoseTrack 2017 and PoseTrack 2018 datasets demonstrate that the proposed method achieves results superior to the state of the art on both human pose estimation and tracking tasks.
arXiv Detail & Related papers (2021-06-07T16:36:50Z) - Learning Uncertainty For Safety-Oriented Semantic Segmentation In
Autonomous Driving [77.39239190539871]
We show how uncertainty estimation can be leveraged to enable safety critical image segmentation in autonomous driving.
We introduce a new uncertainty measure based on disagreeing predictions as measured by a dissimilarity function.
We show experimentally that our proposed approach is much less computationally intensive at inference time than competing methods.
arXiv Detail & Related papers (2021-05-28T09:23:05Z) - Self-Supervision by Prediction for Object Discovery in Videos [62.87145010885044]
In this paper, we use the prediction task as self-supervision and build a novel object-centric model for image sequence representation.
Our framework can be trained without the help of any manual annotation or pretrained network.
Initial experiments confirm that the proposed pipeline is a promising step towards object-centric video prediction.
arXiv Detail & Related papers (2021-03-09T19:14:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.