Single-view 3D Body and Cloth Reconstruction under Complex Poses
- URL: http://arxiv.org/abs/2205.04087v1
- Date: Mon, 9 May 2022 07:34:06 GMT
- Title: Single-view 3D Body and Cloth Reconstruction under Complex Poses
- Authors: Nicolas Ugrinovic, Albert Pumarola, Alberto Sanfeliu and Francesc
Moreno-Noguer
- Abstract summary: We extend existing implicit function-based models to deal with images of humans with arbitrary poses and self-occluded limbs.
We learn an implicit function that maps the input image to a 3D body shape with a low level of detail.
We then learn a displacement map, conditioned on the smoothed surface, which encodes the high-frequency details of the clothes and body.
- Score: 37.86174829271747
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recent advances in 3D human shape reconstruction from single images have
shown impressive results, leveraging on deep networks that model the so-called
implicit function to learn the occupancy status of arbitrarily dense 3D points
in space. However, while current algorithms based on this paradigm, like
PiFuHD, are able to estimate accurate geometry of the human shape and clothes,
they require high-resolution input images and are not able to capture complex
body poses. Most training and evaluation is performed on 1k-resolution images
of humans standing in front of the camera under neutral body poses. In this
paper, we leverage publicly available data to extend existing implicit
function-based models to deal with images of humans that can have arbitrary
poses and self-occluded limbs. We argue that the representation power of the
implicit function is not sufficient to simultaneously model details of the
geometry and of the body pose. We, therefore, propose a coarse-to-fine approach
in which we first learn an implicit function that maps the input image to a 3D
body shape with a low level of detail, but which correctly fits the underlying
human pose, despite its complexity. We then learn a displacement map,
conditioned on the smoothed surface and on the input image, which encodes the
high-frequency details of the clothes and body. In the experimental section, we
show that this coarse-to-fine strategy represents a very good trade-off between
shape detail and pose correctness, comparing favorably to the most recent
state-of-the-art approaches. Our code will be made publicly available.
Related papers
- Learning Visibility for Robust Dense Human Body Estimation [78.37389398573882]
Estimating 3D human pose and shape from 2D images is a crucial yet challenging task.
We learn dense human body estimation that is robust to partial observations.
We obtain pseudo ground-truths of visibility labels from dense UV correspondences and train a neural network to predict visibility along with 3D coordinates.
arXiv Detail & Related papers (2022-08-23T00:01:05Z) - Disentangling 3D Attributes from a Single 2D Image: Human Pose, Shape
and Garment [20.17991487155361]
We focus on the challenging task of extracting disentangled 3D attributes only from 2D image data.
Our method learns an embedding with disentangled latent representations of these three image properties.
We show how an implicit shape loss can benefit the model's ability to recover fine-grained reconstruction details.
arXiv Detail & Related papers (2022-08-05T13:48:43Z) - Pose-NDF: Modeling Human Pose Manifolds with Neural Distance Fields [47.62275563070933]
We present a continuous model for plausible human poses based on neural distance fields (NDFs)
Pose-NDF learns a manifold of plausible poses as the zero level set of a neural implicit function.
It can be used to generate more diverse poses by random sampling and projection than VAE-based methods.
arXiv Detail & Related papers (2022-07-27T21:46:47Z) - Neural-GIF: Neural Generalized Implicit Functions for Animating People
in Clothing [49.32522765356914]
We learn to animate people in clothing as a function of the body pose.
We learn to map every point in the space to a canonical space, where a learned deformation field is applied to model non-rigid effects.
Neural-GIF can be trained on raw 3D scans and reconstructs detailed complex surface geometry and deformations.
arXiv Detail & Related papers (2021-08-19T17:25:16Z) - Detailed Avatar Recovery from Single Image [50.82102098057822]
This paper presents a novel framework to recover emphdetailed avatar from a single image.
We use the deep neural networks to refine the 3D shape in a Hierarchical Mesh Deformation framework.
Our method can restore detailed human body shapes with complete textures beyond skinned models.
arXiv Detail & Related papers (2021-08-06T03:51:26Z) - SCANimate: Weakly Supervised Learning of Skinned Clothed Avatar Networks [54.94737477860082]
We present an end-to-end trainable framework that takes raw 3D scans of a clothed human and turns them into an animatable avatar.
SCANimate does not rely on a customized mesh template or surface mesh registration.
Our method can be applied to pose-aware appearance modeling to generate a fully textured avatar.
arXiv Detail & Related papers (2021-04-07T17:59:58Z) - Neural 3D Clothes Retargeting from a Single Image [91.5030622330039]
We present a method of clothes; generating the potential poses and deformations of a given 3D clothing template model to fit onto a person in a single RGB image.
The problem is fundamentally ill-posed as attaining the ground truth data is impossible, i.e. images of people wearing the different 3D clothing template model model at exact same pose.
We propose a semi-supervised learning framework that validates the physical plausibility of 3D deformation by matching with the prescribed body-to-cloth contact points and clothing to fit onto the unlabeled silhouette.
arXiv Detail & Related papers (2021-01-29T20:50:34Z) - Combining Implicit Function Learning and Parametric Models for 3D Human
Reconstruction [123.62341095156611]
Implicit functions represented as deep learning approximations are powerful for reconstructing 3D surfaces.
Such features are essential in building flexible models for both computer graphics and computer vision.
We present methodology that combines detail-rich implicit functions and parametric representations.
arXiv Detail & Related papers (2020-07-22T13:46:14Z) - Learning 3D Human Shape and Pose from Dense Body Parts [117.46290013548533]
We propose a Decompose-and-aggregate Network (DaNet) to learn 3D human shape and pose from dense correspondences of body parts.
Messages from local streams are aggregated to enhance the robust prediction of the rotation-based poses.
Our method is validated on both indoor and real-world datasets including Human3.6M, UP3D, COCO, and 3DPW.
arXiv Detail & Related papers (2019-12-31T15:09:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.