3D Shape Reconstruction from Vision and Touch
- URL: http://arxiv.org/abs/2007.03778v2
- Date: Mon, 2 Nov 2020 19:57:55 GMT
- Title: 3D Shape Reconstruction from Vision and Touch
- Authors: Edward J. Smith, Roberto Calandra, Adriana Romero, Georgia Gkioxari,
David Meger, Jitendra Malik, Michal Drozdzal
- Abstract summary: In 3D shape reconstruction, the complementary fusion of visual and haptic modalities remains largely unexplored.
We introduce a dataset of simulated touch and vision signals from the interaction between a robotic hand and a large array of 3D objects.
- Score: 62.59044232597045
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When a toddler is presented a new toy, their instinctual behaviour is to pick
it upand inspect it with their hand and eyes in tandem, clearly searching over
its surface to properly understand what they are playing with. At any instance
here, touch provides high fidelity localized information while vision provides
complementary global context. However, in 3D shape reconstruction, the
complementary fusion of visual and haptic modalities remains largely
unexplored. In this paper, we study this problem and present an effective
chart-based approach to multi-modal shape understanding which encourages a
similar fusion vision and touch information.To do so, we introduce a dataset of
simulated touch and vision signals from the interaction between a robotic hand
and a large array of 3D objects. Our results show that (1) leveraging both
vision and touch signals consistently improves single-modality baselines; (2)
our approach outperforms alternative modality fusion methods and strongly
benefits from the proposed chart-based structure; (3) there construction
quality increases with the number of grasps provided; and (4) the touch
information not only enhances the reconstruction at the touch site but also
extrapolates to its local neighborhood.
Related papers
- FusionSense: Bridging Common Sense, Vision, and Touch for Robust Sparse-View Reconstruction [17.367277970910813]
Humans effortlessly integrate common-sense knowledge with sensory input from vision and touch to understand their surroundings.
We introduce FusionSense, a novel 3D reconstruction framework that enables robots to fuse priors from foundation models with highly sparse observations from vision and tactile sensors.
arXiv Detail & Related papers (2024-10-10T18:07:07Z) - Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding [50.448520056844885]
We propose a generative Bayesian network to produce diverse synthetic scenes with real-world patterns.
A series of experiments robustly display our method's consistent superiority over existing state-of-the-art pre-training approaches.
arXiv Detail & Related papers (2024-06-17T07:43:53Z) - Snap-it, Tap-it, Splat-it: Tactile-Informed 3D Gaussian Splatting for Reconstructing Challenging Surfaces [34.831730064258494]
We propose Tactile-Informed 3DGS, a novel approach that incorporates touch data (local depth maps) with multi-view vision data to achieve surface reconstruction and novel view synthesis.
By creating a framework that decreases the transmittance at touch locations, we achieve a refined surface reconstruction, ensuring a uniformly smooth depth map.
We conduct evaluation on objects with glossy and reflective surfaces and demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2024-03-29T16:30:17Z) - Unifying Correspondence, Pose and NeRF for Pose-Free Novel View Synthesis from Stereo Pairs [57.492124844326206]
This work delves into the task of pose-free novel view synthesis from stereo pairs, a challenging and pioneering task in 3D vision.
Our innovative framework, unlike any before, seamlessly integrates 2D correspondence matching, camera pose estimation, and NeRF rendering, fostering a synergistic enhancement of these tasks.
arXiv Detail & Related papers (2023-12-12T13:22:44Z) - Learning Explicit Contact for Implicit Reconstruction of Hand-held
Objects from Monocular Images [59.49985837246644]
We show how to model contacts in an explicit way to benefit the implicit reconstruction of hand-held objects.
In the first part, we propose a new subtask of directly estimating 3D hand-object contacts from a single image.
In the second part, we introduce a novel method to diffuse estimated contact states from the hand mesh surface to nearby 3D space.
arXiv Detail & Related papers (2023-05-31T17:59:26Z) - Collaborative Learning for Hand and Object Reconstruction with
Attention-guided Graph Convolution [49.10497573378427]
Estimating the pose and shape of hands and objects under interaction finds numerous applications including augmented and virtual reality.
Our algorithm is optimisation to object models, and it learns the physical rules governing hand-object interaction.
Experiments using four widely-used benchmarks show that our framework achieves beyond state-of-the-art accuracy in 3D pose estimation, as well as recovers dense 3D hand and object shapes.
arXiv Detail & Related papers (2022-04-27T17:00:54Z) - Active 3D Shape Reconstruction from Vision and Touch [66.08432412497443]
Humans build 3D understandings of the world through active object exploration, using jointly their senses of vision and touch.
In 3D shape reconstruction, most recent progress has relied on static datasets of limited sensory data such as RGB images, depth maps or haptic readings.
We introduce a system composed of: 1) a haptic simulator leveraging high spatial resolution vision-based tactile sensors for active touching of 3D objects; 2) a mesh-based 3D shape reconstruction model that relies on tactile or visuotactile priors to guide the shape exploration; and 3) a set of data-driven solutions with either tactile or visuo
arXiv Detail & Related papers (2021-07-20T15:56:52Z) - Look, Evolve and Mold: Learning 3D Shape Manifold via Single-view
Synthetic Data [32.54820023526409]
We propose a domain-adaptive network for single-view 3D reconstruction, dubbed LEM, to generalize towards the natural scenario.
Experiments on several benchmarks demonstrate the effectiveness and robustness of the proposed method.
arXiv Detail & Related papers (2021-03-08T14:30:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.