CHORD: Category-level Hand-held Object Reconstruction via Shape
Deformation
- URL: http://arxiv.org/abs/2308.10574v1
- Date: Mon, 21 Aug 2023 09:14:18 GMT
- Title: CHORD: Category-level Hand-held Object Reconstruction via Shape
Deformation
- Authors: Kailin Li, Lixin Yang, Haoyu Zhen, Zenan Lin, Xinyu Zhan, Licheng
Zhong, Jian Xu, Kejian Wu, Cewu Lu
- Abstract summary: In daily life, humans utilize hands to manipulate objects.
Previous approaches have encountered difficulties in reconstructing the precise shapes of hand-held objects.
We propose a new method, CHORD, for Category-level Hand-held Object Reconstruction via shape Deformation.
- Score: 40.58622555407404
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In daily life, humans utilize hands to manipulate objects. Modeling the shape
of objects that are manipulated by the hand is essential for AI to comprehend
daily tasks and to learn manipulation skills. However, previous approaches have
encountered difficulties in reconstructing the precise shapes of hand-held
objects, primarily owing to a deficiency in prior shape knowledge and
inadequate data for training. As illustrated, given a particular type of tool,
such as a mug, despite its infinite variations in shape and appearance, humans
have a limited number of 'effective' modes and poses for its manipulation. This
can be attributed to the fact that humans have mastered the shape prior of the
'mug' category, and can quickly establish the corresponding relations between
different mug instances and the prior, such as where the rim and handle are
located. In light of this, we propose a new method, CHORD, for Category-level
Hand-held Object Reconstruction via shape Deformation. CHORD deforms a
categorical shape prior for reconstructing the intra-class objects. To ensure
accurate reconstruction, we empower CHORD with three types of awareness:
appearance, shape, and interacting pose. In addition, we have constructed a new
dataset, COMIC, of category-level hand-object interaction. COMIC contains a
rich array of object instances, materials, hand interactions, and viewing
directions. Extensive evaluation shows that CHORD outperforms state-of-the-art
approaches in both quantitative and qualitative measures. Code, model, and
datasets are available at https://kailinli.github.io/CHORD.
Related papers
- PartHOI: Part-based Hand-Object Interaction Transfer via Generalized Cylinders [15.1049019475729]
Learning-based methods to understand and model hand-object interactions (HOI) require a large amount of high-quality HOI data.
One way to create HOI data is to transfer hand poses from a source object to another based on the objects' geometry.
PartHOI establishes a robust geometric correspondence between object parts, and enables the transfer of contact points.
arXiv Detail & Related papers (2025-04-29T09:56:29Z) - EasyHOI: Unleashing the Power of Large Models for Reconstructing Hand-Object Interactions in the Wild [79.71523320368388]
Our work aims to reconstruct hand-object interactions from a single-view image.
We first design a novel pipeline to estimate the underlying hand pose and object shape.
With the initial reconstruction, we employ a prior-guided optimization scheme.
arXiv Detail & Related papers (2024-11-21T16:33:35Z) - ShapeGraFormer: GraFormer-Based Network for Hand-Object Reconstruction from a Single Depth Map [11.874184782686532]
We propose the first approach for realistic 3D hand-object shape and pose reconstruction from a single depth map.
Our pipeline additionally predicts voxelized hand-object shapes, having a one-to-one mapping to the input voxelized depth.
In addition, we show the impact of adding another GraFormer component that refines the reconstructed shapes based on the hand-object interactions.
arXiv Detail & Related papers (2023-10-18T09:05:57Z) - DTF-Net: Category-Level Pose Estimation and Shape Reconstruction via
Deformable Template Field [29.42222066097076]
Estimating 6D poses and reconstructing 3D shapes of objects in open-world scenes from RGB-depth image pairs is challenging.
We propose the DTF-Net, a novel framework for pose estimation and shape reconstruction based on implicit neural fields of object categories.
arXiv Detail & Related papers (2023-08-04T10:35:40Z) - Learning to Transfer In-Hand Manipulations Using a Greedy Shape
Curriculum [79.6027464700869]
We show that natural and robust in-hand manipulation of simple objects in a dynamic simulation can be learned from a high quality motion capture example.
We propose a simple greedy curriculum search algorithm that can successfully apply to a range of objects such as a teapot, bunny, bottle, train, and elephant.
arXiv Detail & Related papers (2023-03-14T17:08:19Z) - HMDO: Markerless Multi-view Hand Manipulation Capture with Deformable
Objects [8.711239906965893]
HMDO is the first markerless deformable interaction dataset recording interactive motions of the hands and deformable objects.
The proposed method can reconstruct interactive motions of hands and deformable objects with high quality.
arXiv Detail & Related papers (2023-01-18T16:55:15Z) - Continuous Surface Embeddings [76.86259029442624]
We focus on the task of learning and representing dense correspondences in deformable object categories.
We propose a new, learnable image-based representation of dense correspondences.
We demonstrate that the proposed approach performs on par or better than the state-of-the-art methods for dense pose estimation for humans.
arXiv Detail & Related papers (2020-11-24T22:52:15Z) - Unsupervised Shape and Pose Disentanglement for 3D Meshes [49.431680543840706]
We present a simple yet effective approach to learn disentangled shape and pose representations in an unsupervised setting.
We use a combination of self-consistency and cross-consistency constraints to learn pose and shape space from registered meshes.
We demonstrate the usefulness of learned representations through a number of tasks including pose transfer and shape retrieval.
arXiv Detail & Related papers (2020-07-22T11:00:27Z) - Shape Prior Deformation for Categorical 6D Object Pose and Size
Estimation [62.618227434286]
We present a novel learning approach to recover the 6D poses and sizes of unseen object instances from an RGB-D image.
We propose a deep network to reconstruct the 3D object model by explicitly modeling the deformation from a pre-learned categorical shape prior.
arXiv Detail & Related papers (2020-07-16T16:45:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.