PaMIR: Parametric Model-Conditioned Implicit Representation for
Image-based Human Reconstruction
- URL: http://arxiv.org/abs/2007.03858v2
- Date: Tue, 15 Dec 2020 12:12:03 GMT
- Title: PaMIR: Parametric Model-Conditioned Implicit Representation for
Image-based Human Reconstruction
- Authors: Zerong Zheng and Tao Yu and Yebin Liu and Qionghai Dai
- Abstract summary: We propose Parametric Model-Conditioned Implicit Representation (PaMIR), which combines the parametric body model with the free-form deep implicit function.
We show that our method achieves state-of-the-art performance for image-based 3D human reconstruction in the cases of challenging poses and clothing types.
- Score: 67.08350202974434
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modeling 3D humans accurately and robustly from a single image is very
challenging, and the key for such an ill-posed problem is the 3D representation
of the human models. To overcome the limitations of regular 3D representations,
we propose Parametric Model-Conditioned Implicit Representation (PaMIR), which
combines the parametric body model with the free-form deep implicit function.
In our PaMIR-based reconstruction framework, a novel deep neural network is
proposed to regularize the free-form deep implicit function using the semantic
features of the parametric model, which improves the generalization ability
under the scenarios of challenging poses and various clothing topologies.
Moreover, a novel depth-ambiguity-aware training loss is further integrated to
resolve depth ambiguities and enable successful surface detail reconstruction
with imperfect body reference. Finally, we propose a body reference
optimization method to improve the parametric model estimation accuracy and to
enhance the consistency between the parametric model and the implicit function.
With the PaMIR representation, our framework can be easily extended to
multi-image input scenarios without the need of multi-camera calibration and
pose synchronization. Experimental results demonstrate that our method achieves
state-of-the-art performance for image-based 3D human reconstruction in the
cases of challenging poses and clothing types.
Related papers
- Towards Robust and Expressive Whole-body Human Pose and Shape Estimation [51.457517178632756]
Whole-body pose and shape estimation aims to jointly predict different behaviors of the entire human body from a monocular image.
Existing methods often exhibit degraded performance under the complexity of in-the-wild scenarios.
We propose a novel framework to enhance the robustness of whole-body pose and shape estimation.
arXiv Detail & Related papers (2023-12-14T08:17:42Z) - HAvatar: High-fidelity Head Avatar via Facial Model Conditioned Neural
Radiance Field [44.848368616444446]
We introduce a novel hybrid explicit-implicit 3D representation, Facial Model Conditioned Neural Radiance Field, which integrates the expressiveness of NeRF and the prior information from the parametric template.
By adopting an overall GAN-based architecture using an image-to-image translation network, we achieve high-resolution, realistic and view-consistent synthesis of dynamic head appearance.
arXiv Detail & Related papers (2023-09-29T10:45:22Z) - FrozenRecon: Pose-free 3D Scene Reconstruction with Frozen Depth Models [67.96827539201071]
We propose a novel test-time optimization approach for 3D scene reconstruction.
Our method achieves state-of-the-art cross-dataset reconstruction on five zero-shot testing datasets.
arXiv Detail & Related papers (2023-08-10T17:55:02Z) - MA-NeRF: Motion-Assisted Neural Radiance Fields for Face Synthesis from
Sparse Images [21.811067296567252]
We propose a novel framework that can reconstruct a high-fidelity drivable face avatar and handle unseen expressions.
At the core of our implementation are structured displacement feature and semantic-aware learning module.
Our method achieves much better results than the current state-of-the-arts.
arXiv Detail & Related papers (2023-06-17T13:49:56Z) - A Probabilistic Attention Model with Occlusion-aware Texture Regression
for 3D Hand Reconstruction from a Single RGB Image [5.725477071353354]
Deep learning approaches have shown promising results in 3D hand reconstruction from a single RGB image.
We propose a novel probabilistic model to achieve the robustness of model-based approaches.
We demonstrate the flexibility of the proposed probabilistic model to be trained in both supervised and weakly-supervised scenarios.
arXiv Detail & Related papers (2023-04-27T16:02:32Z) - ASM: Adaptive Skinning Model for High-Quality 3D Face Modeling [11.885382595302751]
We argue that reconstruction with multi-view uncalibrated images demands a new model with stronger capacity.
We propose Adaptive Skinning Model (ASM), which redefines the skinning model with more compact and fully tunable parameters.
Our work opens up new research direction for parametric face model and facilitates future research on multi-view reconstruction.
arXiv Detail & Related papers (2023-04-19T04:55:28Z) - Adversarial Parametric Pose Prior [106.12437086990853]
We learn a prior that restricts the SMPL parameters to values that produce realistic poses via adversarial training.
We show that our learned prior covers the diversity of the real-data distribution, facilitates optimization for 3D reconstruction from 2D keypoints, and yields better pose estimates when used for regression from images.
arXiv Detail & Related papers (2021-12-08T10:05:32Z) - LatentHuman: Shape-and-Pose Disentangled Latent Representation for Human
Bodies [78.17425779503047]
We propose a novel neural implicit representation for the human body.
It is fully differentiable and optimizable with disentangled shape and pose latent spaces.
Our model can be trained and fine-tuned directly on non-watertight raw data with well-designed losses.
arXiv Detail & Related papers (2021-11-30T04:10:57Z) - Neural Descent for Visual 3D Human Pose and Shape [67.01050349629053]
We present deep neural network methodology to reconstruct the 3d pose and shape of people, given an input RGB image.
We rely on a recently introduced, expressivefull body statistical 3d human model, GHUM, trained end-to-end.
Central to our methodology, is a learning to learn and optimize approach, referred to as HUmanNeural Descent (HUND), which avoids both second-order differentiation.
arXiv Detail & Related papers (2020-08-16T13:38:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.