Contact2Grasp: 3D Grasp Synthesis via Hand-Object Contact Constraint
- URL: http://arxiv.org/abs/2210.09245v3
- Date: Sat, 6 May 2023 07:53:13 GMT
- Title: Contact2Grasp: 3D Grasp Synthesis via Hand-Object Contact Constraint
- Authors: Haoming Li, Xinzhuo Lin, Yang Zhou, Xiang Li, Yuchi Huo, Jiming Chen
and Qi Ye
- Abstract summary: 3D grasp synthesis generates grasping poses given an input object.
We introduce an intermediate variable for grasp contact areas to constrain the grasp generation.
Our method outperforms state-of-the-art methods regarding grasp generation on various metrics.
- Score: 18.201389966034263
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D grasp synthesis generates grasping poses given an input object. Existing
works tackle the problem by learning a direct mapping from objects to the
distributions of grasping poses. However, because the physical contact is
sensitive to small changes in pose, the high-nonlinear mapping between 3D
object representation to valid poses is considerably non-smooth, leading to
poor generation efficiency and restricted generality. To tackle the challenge,
we introduce an intermediate variable for grasp contact areas to constrain the
grasp generation; in other words, we factorize the mapping into two sequential
stages by assuming that grasping poses are fully constrained given contact
maps: 1) we first learn contact map distributions to generate the potential
contact maps for grasps; 2) then learn a mapping from the contact maps to the
grasping poses. Further, we propose a penetration-aware optimization with the
generated contacts as a consistency constraint for grasp refinement. Extensive
validations on two public datasets show that our method outperforms
state-of-the-art methods regarding grasp generation on various metrics.
Related papers
- ClickDiff: Click to Induce Semantic Contact Map for Controllable Grasp Generation with Diffusion Models [17.438429495623755]
ClickDiff is a controllable conditional generation model that leverages a fine-grained Semantic Contact Map.
Within this framework, the Semantic Conditional Module generates reasonable contact maps based on fine-grained contact information.
We evaluate the validity of our proposed method, demonstrating the efficacy and robustness of ClickDiff, even with previously unseen objects.
arXiv Detail & Related papers (2024-07-28T02:42:29Z) - Text2HOI: Text-guided 3D Motion Generation for Hand-Object Interaction [8.253265795150401]
This paper introduces the first text-guided work for generating the sequence of hand-object interaction in 3D.
For contact generation, a VAE-based network takes as input a text and an object mesh, and generates the probability of contacts between the surfaces of hands and the object.
For motion generation, a Transformer-based diffusion model utilizes this 3D contact map as a strong prior for generating physically plausible hand-object motion.
arXiv Detail & Related papers (2024-03-31T04:56:30Z) - Neural Semantic Surface Maps [52.61017226479506]
We present an automated technique for computing a map between two genus-zero shapes, which matches semantically corresponding regions to one another.
Our approach can generate semantic surface-to-surface maps, eliminating manual annotations or any 3D training data requirement.
arXiv Detail & Related papers (2023-09-09T16:21:56Z) - Learning Explicit Contact for Implicit Reconstruction of Hand-held
Objects from Monocular Images [59.49985837246644]
We show how to model contacts in an explicit way to benefit the implicit reconstruction of hand-held objects.
In the first part, we propose a new subtask of directly estimating 3D hand-object contacts from a single image.
In the second part, we introduce a novel method to diffuse estimated contact states from the hand mesh surface to nearby 3D space.
arXiv Detail & Related papers (2023-05-31T17:59:26Z) - S$^2$Contact: Graph-based Network for 3D Hand-Object Contact Estimation
with Semi-Supervised Learning [70.72037296392642]
We propose a novel semi-supervised framework that allows us to learn contact from monocular images.
Specifically, we leverage visual and geometric consistency constraints in large-scale datasets for generating pseudo-labels.
We show benefits from using a contact map that rules hand-object interactions to produce more accurate reconstructions.
arXiv Detail & Related papers (2022-08-01T14:05:23Z) - Collaborative Learning for Hand and Object Reconstruction with
Attention-guided Graph Convolution [49.10497573378427]
Estimating the pose and shape of hands and objects under interaction finds numerous applications including augmented and virtual reality.
Our algorithm is optimisation to object models, and it learns the physical rules governing hand-object interaction.
Experiments using four widely-used benchmarks show that our framework achieves beyond state-of-the-art accuracy in 3D pose estimation, as well as recovers dense 3D hand and object shapes.
arXiv Detail & Related papers (2022-04-27T17:00:54Z) - Non-Local Latent Relation Distillation for Self-Adaptive 3D Human Pose
Estimation [63.199549837604444]
3D human pose estimation approaches leverage different forms of strong (2D/3D pose) or weak (multi-view or depth) paired supervision.
We cast 3D pose learning as a self-supervised adaptation problem that aims to transfer the task knowledge from a labeled source domain to a completely unpaired target.
We evaluate different self-adaptation settings and demonstrate state-of-the-art 3D human pose estimation performance on standard benchmarks.
arXiv Detail & Related papers (2022-04-05T03:52:57Z) - 3D Shape Reconstruction from Vision and Touch [62.59044232597045]
In 3D shape reconstruction, the complementary fusion of visual and haptic modalities remains largely unexplored.
We introduce a dataset of simulated touch and vision signals from the interaction between a robotic hand and a large array of 3D objects.
arXiv Detail & Related papers (2020-07-07T20:20:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.