Leveraging Photometric Consistency over Time for Sparsely Supervised
Hand-Object Reconstruction
- URL: http://arxiv.org/abs/2004.13449v1
- Date: Tue, 28 Apr 2020 12:03:14 GMT
- Title: Leveraging Photometric Consistency over Time for Sparsely Supervised
Hand-Object Reconstruction
- Authors: Yana Hasson, Bugra Tekin, Federica Bogo, Ivan Laptev, Marc Pollefeys,
Cordelia Schmid
- Abstract summary: We present a method to leverage photometric consistency across time when annotations are only available for a sparse subset of frames in a video.
Our model is trained end-to-end on color images to jointly reconstruct hands and objects in 3D by inferring their poses.
We achieve state-of-the-art results on 3D hand-object reconstruction benchmarks and demonstrate that our approach allows us to improve the pose estimation accuracy.
- Score: 118.21363599332493
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modeling hand-object manipulations is essential for understanding how humans
interact with their environment. While of practical importance, estimating the
pose of hands and objects during interactions is challenging due to the large
mutual occlusions that occur during manipulation. Recent efforts have been
directed towards fully-supervised methods that require large amounts of labeled
training samples. Collecting 3D ground-truth data for hand-object interactions,
however, is costly, tedious, and error-prone. To overcome this challenge we
present a method to leverage photometric consistency across time when
annotations are only available for a sparse subset of frames in a video. Our
model is trained end-to-end on color images to jointly reconstruct hands and
objects in 3D by inferring their poses. Given our estimated reconstructions, we
differentiably render the optical flow between pairs of adjacent images and use
it within the network to warp one frame to another. We then apply a
self-supervised photometric loss that relies on the visual consistency between
nearby images. We achieve state-of-the-art results on 3D hand-object
reconstruction benchmarks and demonstrate that our approach allows us to
improve the pose estimation accuracy by leveraging information from neighboring
frames in low-data regimes.
Related papers
- EasyHOI: Unleashing the Power of Large Models for Reconstructing Hand-Object Interactions in the Wild [79.71523320368388]
Our work aims to reconstruct hand-object interactions from a single-view image.
We first design a novel pipeline to estimate the underlying hand pose and object shape.
With the initial reconstruction, we employ a prior-guided optimization scheme.
arXiv Detail & Related papers (2024-11-21T16:33:35Z) - DICE: End-to-end Deformation Capture of Hand-Face Interactions from a Single Image [98.29284902879652]
We present DICE, the first end-to-end method for Deformation-aware hand-face Interaction reCovEry from a single image.
It features disentangling the regression of local deformation fields and global mesh locations into two network branches.
It achieves state-of-the-art performance on a standard benchmark and in-the-wild data in terms of accuracy and physical plausibility.
arXiv Detail & Related papers (2024-06-26T00:08:29Z) - Denoising Diffusion for 3D Hand Pose Estimation from Images [38.20064386142944]
This paper addresses the problem of 3D hand pose estimation from monocular images or sequences.
We present a novel end-to-end framework for 3D hand regression that employs diffusion models that have shown excellent ability to capture the distribution of data for generative purposes.
The proposed model provides state-of-the-art performance when lifting a 2D single-hand image to 3D.
arXiv Detail & Related papers (2023-08-18T12:57:22Z) - TOCH: Spatio-Temporal Object-to-Hand Correspondence for Motion
Refinement [42.3418874174372]
We present TOCH, a method for refining incorrect 3D hand-object interaction sequences using a data prior.
We learn a latent manifold of plausible TOCH fields with a temporal denoising auto-encoder.
Experiments demonstrate that TOCH outperforms state-of-the-art 3D hand-object interaction models.
arXiv Detail & Related papers (2022-05-16T20:41:45Z) - What's in your hands? 3D Reconstruction of Generic Objects in Hands [49.12461675219253]
Our work aims to reconstruct hand-held objects given a single RGB image.
In contrast to prior works that typically assume known 3D templates and reduce the problem to 3D pose estimation, our work reconstructs generic hand-held object without knowing their 3D templates.
arXiv Detail & Related papers (2022-04-14T17:59:02Z) - Semi-Supervised 3D Hand-Object Poses Estimation with Interactions in
Time [22.574069344246052]
We propose a unified framework for estimating the 3D hand and object poses with semi-supervised learning.
We build a joint learning framework where we perform explicit contextual reasoning between hand and object representations by a Transformer.
Our method not only improves hand pose estimation in challenging real-world dataset, but also substantially improve the object pose which has fewer ground-truths per instance.
arXiv Detail & Related papers (2021-06-09T17:59:34Z) - SeqHAND:RGB-Sequence-Based 3D Hand Pose and Shape Estimation [48.456638103309544]
3D hand pose estimation based on RGB images has been studied for a long time.
We propose a novel method that generates a synthetic dataset that mimics natural human hand movements.
We show that utilizing temporal information for 3D hand pose estimation significantly enhances general pose estimations.
arXiv Detail & Related papers (2020-07-10T05:11:14Z) - Measuring Generalisation to Unseen Viewpoints, Articulations, Shapes and
Objects for 3D Hand Pose Estimation under Hand-Object Interaction [137.28465645405655]
HANDS'19 is a challenge to evaluate the abilities of current 3D hand pose estimators (HPEs) to interpolate and extrapolate the poses of a training set.
We show that the accuracy of state-of-the-art methods can drop, and that they fail mostly on poses absent from the training set.
arXiv Detail & Related papers (2020-03-30T19:28:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.