Implicit 3D Human Mesh Recovery using Consistency with Pose and Shape
from Unseen-view
- URL: http://arxiv.org/abs/2306.17651v2
- Date: Mon, 3 Jul 2023 01:29:19 GMT
- Title: Implicit 3D Human Mesh Recovery using Consistency with Pose and Shape
from Unseen-view
- Authors: Hanbyel Cho, Yooshin Cho, Jaesung Ahn, Junmo Kim
- Abstract summary: From an image of a person, we can easily infer the natural 3D pose and shape of the person even if ambiguity exists.
We propose "Implicit 3D Human Mesh Recovery (ImpHMR)" that can implicitly imagine a person in 3D space at the feature-level.
- Score: 23.30144908939099
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: From an image of a person, we can easily infer the natural 3D pose and shape
of the person even if ambiguity exists. This is because we have a mental model
that allows us to imagine a person's appearance at different viewing directions
from a given image and utilize the consistency between them for inference.
However, existing human mesh recovery methods only consider the direction in
which the image was taken due to their structural limitations. Hence, we
propose "Implicit 3D Human Mesh Recovery (ImpHMR)" that can implicitly imagine
a person in 3D space at the feature-level via Neural Feature Fields. In ImpHMR,
feature fields are generated by CNN-based image encoder for a given image.
Then, the 2D feature map is volume-rendered from the feature field for a given
viewing direction, and the pose and shape parameters are regressed from the
feature. To utilize consistency with pose and shape from unseen-view, if there
are 3D labels, the model predicts results including the silhouette from an
arbitrary direction and makes it equal to the rotated ground-truth. In the case
of only 2D labels, we perform self-supervised learning through the constraint
that the pose and shape parameters inferred from different directions should be
the same. Extensive evaluations show the efficacy of the proposed method.
Related papers
- Disentangling 3D Attributes from a Single 2D Image: Human Pose, Shape
and Garment [20.17991487155361]
We focus on the challenging task of extracting disentangled 3D attributes only from 2D image data.
Our method learns an embedding with disentangled latent representations of these three image properties.
We show how an implicit shape loss can benefit the model's ability to recover fine-grained reconstruction details.
arXiv Detail & Related papers (2022-08-05T13:48:43Z) - Single-view 3D Body and Cloth Reconstruction under Complex Poses [37.86174829271747]
We extend existing implicit function-based models to deal with images of humans with arbitrary poses and self-occluded limbs.
We learn an implicit function that maps the input image to a 3D body shape with a low level of detail.
We then learn a displacement map, conditioned on the smoothed surface, which encodes the high-frequency details of the clothes and body.
arXiv Detail & Related papers (2022-05-09T07:34:06Z) - Beyond 3DMM: Learning to Capture High-fidelity 3D Face Shape [77.95154911528365]
3D Morphable Model (3DMM) fitting has widely benefited face analysis due to its strong 3D priori.
Previous reconstructed 3D faces suffer from degraded visual verisimilitude due to the loss of fine-grained geometry.
This paper proposes a complete solution to capture the personalized shape so that the reconstructed shape looks identical to the corresponding person.
arXiv Detail & Related papers (2022-04-09T03:46:18Z) - Pop-Out Motion: 3D-Aware Image Deformation via Learning the Shape
Laplacian [58.704089101826774]
We present a 3D-aware image deformation method with minimal restrictions on shape category and deformation type.
We take a supervised learning-based approach to predict the shape Laplacian of the underlying volume of a 3D reconstruction represented as a point cloud.
In the experiments, we present our results of deforming 2D character and clothed human images.
arXiv Detail & Related papers (2022-03-29T04:57:18Z) - Detailed Avatar Recovery from Single Image [50.82102098057822]
This paper presents a novel framework to recover emphdetailed avatar from a single image.
We use the deep neural networks to refine the 3D shape in a Hierarchical Mesh Deformation framework.
Our method can restore detailed human body shapes with complete textures beyond skinned models.
arXiv Detail & Related papers (2021-08-06T03:51:26Z) - SNARF: Differentiable Forward Skinning for Animating Non-Rigid Neural
Implicit Shapes [117.76767853430243]
We introduce SNARF, which combines the advantages of linear blend skinning for polygonal meshes with neural implicit surfaces.
We propose a forward skinning model that finds all canonical correspondences of any deformed point using iterative root finding.
Compared to state-of-the-art neural implicit representations, our approach generalizes better to unseen poses while preserving accuracy.
arXiv Detail & Related papers (2021-04-08T17:54:59Z) - Pose2Mesh: Graph Convolutional Network for 3D Human Pose and Mesh
Recovery from a 2D Human Pose [70.23652933572647]
We propose a novel graph convolutional neural network (GraphCNN)-based system that estimates the 3D coordinates of human mesh vertices directly from the 2D human pose.
We show that our Pose2Mesh outperforms the previous 3D human pose and mesh estimation methods on various benchmark datasets.
arXiv Detail & Related papers (2020-08-20T16:01:56Z) - Chained Representation Cycling: Learning to Estimate 3D Human Pose and
Shape by Cycling Between Representations [73.11883464562895]
We propose a new architecture that facilitates unsupervised, or lightly supervised, learning.
We demonstrate the method by learning 3D human pose and shape from un-paired and un-annotated images.
While we present results for modeling humans, our formulation is general and can be applied to other vision problems.
arXiv Detail & Related papers (2020-01-06T14:54:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.