PoseVocab: Learning Joint-structured Pose Embeddings for Human Avatar
Modeling
- URL: http://arxiv.org/abs/2304.13006v2
- Date: Sun, 14 May 2023 13:15:20 GMT
- Title: PoseVocab: Learning Joint-structured Pose Embeddings for Human Avatar
Modeling
- Authors: Zhe Li, Zerong Zheng, Yuxiao Liu, Boyao Zhou, Yebin Liu
- Abstract summary: We present PoseVocab, a novel pose encoding method that can encode high-fidelity human details.
Given multi-view RGB videos of a character, PoseVocab constructs key poses and latent embeddings based on the training poses.
Experiments show that our method outperforms other state-of-the-art baselines.
- Score: 30.93155530590843
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Creating pose-driven human avatars is about modeling the mapping from the
low-frequency driving pose to high-frequency dynamic human appearances, so an
effective pose encoding method that can encode high-fidelity human details is
essential to human avatar modeling. To this end, we present PoseVocab, a novel
pose encoding method that encourages the network to discover the optimal pose
embeddings for learning the dynamic human appearance. Given multi-view RGB
videos of a character, PoseVocab constructs key poses and latent embeddings
based on the training poses. To achieve pose generalization and temporal
consistency, we sample key rotations in $so(3)$ of each joint rather than the
global pose vectors, and assign a pose embedding to each sampled key rotation.
These joint-structured pose embeddings not only encode the dynamic appearances
under different key poses, but also factorize the global pose embedding into
joint-structured ones to better learn the appearance variation related to the
motion of each joint. To improve the representation ability of the pose
embedding while maintaining memory efficiency, we introduce feature lines, a
compact yet effective 3D representation, to model more fine-grained details of
human appearances. Furthermore, given a query pose and a spatial position, a
hierarchical query strategy is introduced to interpolate pose embeddings and
acquire the conditional pose feature for dynamic human synthesis. Overall,
PoseVocab effectively encodes the dynamic details of human appearance and
enables realistic and generalized animation under novel poses. Experiments show
that our method outperforms other state-of-the-art baselines both qualitatively
and quantitatively in terms of synthesis quality. Code is available at
https://github.com/lizhe00/PoseVocab.
Related papers
- VINECS: Video-based Neural Character Skinning [82.39776643541383]
We propose a fully automated approach for creating a fully rigged character with pose-dependent skinning weights.
We show that our approach outperforms state-of-the-art while not relying on dense 4D scans.
arXiv Detail & Related papers (2023-07-03T08:35:53Z) - Pose Guided Human Image Synthesis with Partially Decoupled GAN [25.800174118151638]
Pose Guided Human Image Synthesis (PGHIS) is a challenging task of transforming a human image from the reference pose to a target pose.
We propose a method by decoupling the human body into several parts to guide the synthesis of a realistic image of the person.
In addition, we design a multi-head attention-based module for PGHIS.
arXiv Detail & Related papers (2022-10-07T15:31:37Z) - Neural Novel Actor: Learning a Generalized Animatable Neural
Representation for Human Actors [98.24047528960406]
We propose a new method for learning a generalized animatable neural representation from a sparse set of multi-view imagery of multiple persons.
The learned representation can be used to synthesize novel view images of an arbitrary person from a sparse set of cameras, and further animate them with the user's pose control.
arXiv Detail & Related papers (2022-08-25T07:36:46Z) - Single-view 3D Body and Cloth Reconstruction under Complex Poses [37.86174829271747]
We extend existing implicit function-based models to deal with images of humans with arbitrary poses and self-occluded limbs.
We learn an implicit function that maps the input image to a 3D body shape with a low level of detail.
We then learn a displacement map, conditioned on the smoothed surface, which encodes the high-frequency details of the clothes and body.
arXiv Detail & Related papers (2022-05-09T07:34:06Z) - Neural Rendering of Humans in Novel View and Pose from Monocular Video [68.37767099240236]
We introduce a new method that generates photo-realistic humans under novel views and poses given a monocular video as input.
Our method significantly outperforms existing approaches under unseen poses and novel views given monocular videos as input.
arXiv Detail & Related papers (2022-04-04T03:09:20Z) - Neural Actor: Neural Free-view Synthesis of Human Actors with Pose
Control [80.79820002330457]
We propose a new method for high-quality synthesis of humans from arbitrary viewpoints and under arbitrary controllable poses.
Our method achieves better quality than the state-of-the-arts on playback as well as novel pose synthesis, and can even generalize well to new poses that starkly differ from the training poses.
arXiv Detail & Related papers (2021-06-03T17:40:48Z) - SCANimate: Weakly Supervised Learning of Skinned Clothed Avatar Networks [54.94737477860082]
We present an end-to-end trainable framework that takes raw 3D scans of a clothed human and turns them into an animatable avatar.
SCANimate does not rely on a customized mesh template or surface mesh registration.
Our method can be applied to pose-aware appearance modeling to generate a fully textured avatar.
arXiv Detail & Related papers (2021-04-07T17:59:58Z) - Unsupervised 3D Human Pose Representation with Viewpoint and Pose
Disentanglement [63.853412753242615]
Learning a good 3D human pose representation is important for human pose related tasks.
We propose a novel Siamese denoising autoencoder to learn a 3D pose representation.
Our approach achieves state-of-the-art performance on two inherently different tasks.
arXiv Detail & Related papers (2020-07-14T14:25:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.