Collision-aware In-hand 6D Object Pose Estimation using Multiple
Vision-based Tactile Sensors
- URL: http://arxiv.org/abs/2301.13667v1
- Date: Tue, 31 Jan 2023 14:35:26 GMT
- Title: Collision-aware In-hand 6D Object Pose Estimation using Multiple
Vision-based Tactile Sensors
- Authors: Gabriele M. Caddeo, Nicola A. Piga, Fabrizio Bottarel and Lorenzo
Natale
- Abstract summary: We reason on the possible spatial configurations of the sensors along the object surface.
We use selected sensors configurations to optimize over the space of 6D poses.
We rank the obtained poses by penalizing those that are in collision with the sensors.
- Score: 4.886250215151643
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we address the problem of estimating the in-hand 6D pose of an
object in contact with multiple vision-based tactile sensors. We reason on the
possible spatial configurations of the sensors along the object surface.
Specifically, we filter contact hypotheses using geometric reasoning and a
Convolutional Neural Network (CNN), trained on simulated object-agnostic
images, to promote those that better comply with the actual tactile images from
the sensors. We use the selected sensors configurations to optimize over the
space of 6D poses using a Gradient Descent-based approach. We finally rank the
obtained poses by penalizing those that are in collision with the sensors. We
carry out experiments in simulation using the DIGIT vision-based sensor with
several objects, from the standard YCB model set. The results demonstrate that
our approach estimates object poses that are compatible with actual
object-sensor contacts in $87.5\%$ of cases while reaching an average
positional error in the order of $2$ centimeters. Our analysis also includes
qualitative results of experiments with a real DIGIT sensor.
Related papers
- MSSIDD: A Benchmark for Multi-Sensor Denoising [55.41612200877861]
We introduce a new benchmark, the Multi-Sensor SIDD dataset, which is the first raw-domain dataset designed to evaluate the sensor transferability of denoising models.
We propose a sensor consistency training framework that enables denoising models to learn the sensor-invariant features.
arXiv Detail & Related papers (2024-11-18T13:32:59Z) - FeelAnyForce: Estimating Contact Force Feedback from Tactile Sensation for Vision-Based Tactile Sensors [18.88211706267447]
We tackle the problem of estimating 3D contact forces using vision-based tactile sensors.
Our goal is to estimate contact forces over a large range (up to 15 N) on any objects while generalizing across different vision-based tactile sensors.
arXiv Detail & Related papers (2024-10-02T21:28:19Z) - On the Importance of Accurate Geometry Data for Dense 3D Vision Tasks [61.74608497496841]
Training on inaccurate or corrupt data induces model bias and hampers generalisation capabilities.
This paper investigates the effect of sensor errors for the dense 3D vision tasks of depth estimation and reconstruction.
arXiv Detail & Related papers (2023-03-26T22:32:44Z) - Tactile-Filter: Interactive Tactile Perception for Part Mating [54.46221808805662]
Humans rely on touch and tactile sensing for a lot of dexterous manipulation tasks.
vision-based tactile sensors are being widely used for various robotic perception and control tasks.
We present a method for interactive perception using vision-based tactile sensors for a part mating task.
arXiv Detail & Related papers (2023-03-10T16:27:37Z) - Monocular 3D Object Detection with Depth from Motion [74.29588921594853]
We take advantage of camera ego-motion for accurate object depth estimation and detection.
Our framework, named Depth from Motion (DfM), then uses the established geometry to lift 2D image features to the 3D space and detects 3D objects thereon.
Our framework outperforms state-of-the-art methods by a large margin on the KITTI benchmark.
arXiv Detail & Related papers (2022-07-26T15:48:46Z) - Coupled Iterative Refinement for 6D Multi-Object Pose Estimation [64.7198752089041]
Given a set of known 3D objects and an RGB or RGB-D input image, we detect and estimate the 6D pose of each object.
Our approach iteratively refines both pose and correspondence in a tightly coupled manner, allowing us to dynamically remove outliers to improve accuracy.
arXiv Detail & Related papers (2022-04-26T18:00:08Z) - Tac2Pose: Tactile Object Pose Estimation from the First Touch [6.321662423735226]
We present Tac2Pose, an object-specific approach to tactile pose estimation from the first touch for known objects.
We simulate the contact shapes that a dense set of object poses would produce on the sensor.
We obtain contact shapes from the sensor with an object-agnostic calibration step that maps RGB tactile observations to binary contact shapes.
arXiv Detail & Related papers (2022-04-25T14:43:48Z) - Monocular Depth Estimation for Soft Visuotactile Sensors [24.319343057803973]
We investigate the application of state-of-the-art monocular depth estimation to infer dense internal (tactile) depth maps directly from an internal single small IR imaging sensor.
We show that deep networks typically used for long-range depth estimation (1-100m) can be effectively trained for precise predictions at a much shorter range (1-100mm) inside a mostly textureless deformable fluid-filled sensor.
We propose a simple supervised learning process to train an object-agnostic network requiring less than 10 random poses in contact for less than 10 seconds for a small set of diverse objects.
arXiv Detail & Related papers (2021-01-05T17:51:11Z) - Tactile Object Pose Estimation from the First Touch with Geometric
Contact Rendering [19.69677059281393]
We present an approach to tactile pose estimation from the first touch for known objects.
We create an object-agnostic map from real tactile observations to contact shapes.
For a new object with known geometry, we learn a tailored perception model completely in simulation.
arXiv Detail & Related papers (2020-12-09T18:00:35Z) - Deep Soft Procrustes for Markerless Volumetric Sensor Alignment [81.13055566952221]
In this work, we improve markerless data-driven correspondence estimation to achieve more robust multi-sensor spatial alignment.
We incorporate geometric constraints in an end-to-end manner into a typical segmentation based model and bridge the intermediate dense classification task with the targeted pose estimation one.
Our model is experimentally shown to achieve similar results with marker-based methods and outperform the markerless ones, while also being robust to the pose variations of the calibration structure.
arXiv Detail & Related papers (2020-03-23T10:51:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.