High-fidelity 3D Human Digitization from Single 2K Resolution Images
- URL: http://arxiv.org/abs/2303.15108v1
- Date: Mon, 27 Mar 2023 11:22:54 GMT
- Title: High-fidelity 3D Human Digitization from Single 2K Resolution Images
- Authors: Sang-Hun Han, Min-Gyu Park, Ju Hong Yoon, Ju-Mi Kang, Young-Jae Park
and Hae-Gon Jeon
- Abstract summary: We propose 2K2K, which constructs a large-scale 2K human dataset and infers 3D human models from 2K resolution images.
We also provide 2,050 3D human models, including texture maps, 3D joints, and SMPL parameters for research purposes.
- Score: 16.29087820634057
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: High-quality 3D human body reconstruction requires high-fidelity and
large-scale training data and appropriate network design that effectively
exploits the high-resolution input images. To tackle these problems, we propose
a simple yet effective 3D human digitization method called 2K2K, which
constructs a large-scale 2K human dataset and infers 3D human models from 2K
resolution images. The proposed method separately recovers the global shape of
a human and its details. The low-resolution depth network predicts the global
structure from a low-resolution image, and the part-wise image-to-normal
network predicts the details of the 3D human body structure. The
high-resolution depth network merges the global 3D shape and the detailed
structures to infer the high-resolution front and back side depth maps.
Finally, an off-the-shelf mesh generator reconstructs the full 3D human model,
which are available at https://github.com/SangHunHan92/2K2K. In addition, we
also provide 2,050 3D human models, including texture maps, 3D joints, and SMPL
parameters for research purposes. In experiments, we demonstrate competitive
performance over the recent works on various datasets.
Related papers
- FAMOUS: High-Fidelity Monocular 3D Human Digitization Using View Synthesis [51.193297565630886]
The challenge of accurately inferring texture remains, particularly in obscured areas such as the back of a person in frontal-view images.
This limitation in texture prediction largely stems from the scarcity of large-scale and diverse 3D datasets.
We propose leveraging extensive 2D fashion datasets to enhance both texture and shape prediction in 3D human digitization.
arXiv Detail & Related papers (2024-10-13T01:25:05Z) - StructLDM: Structured Latent Diffusion for 3D Human Generation [45.51684124904457]
StructLDM is a diffusion-based unconditional 3D human generative model learned from 2D images.
It enables different levels of controllable 3D human generation and editing, including pose/view/shape control, and high-level tasks including compositional generations, part-aware clothing editing, 3D virtual try-on, etc.
arXiv Detail & Related papers (2024-04-01T17:00:18Z) - Joint2Human: High-quality 3D Human Generation via Compact Spherical Embedding of 3D Joints [8.704552731994452]
Joint2Human is a novel method that leverages 2D diffusion models to generate detailed 3D human geometry directly.
Our method can ensure global structure, local details, high resolution, and low computational cost simultaneously.
arXiv Detail & Related papers (2023-12-14T01:24:22Z) - SHERF: Generalizable Human NeRF from a Single Image [59.10589479808622]
SHERF is the first generalizable Human NeRF model for recovering animatable 3D humans from a single input image.
We propose a bank of 3D-aware hierarchical features, including global, point-level, and pixel-aligned features, to facilitate informative encoding.
arXiv Detail & Related papers (2023-03-22T17:59:12Z) - Get3DHuman: Lifting StyleGAN-Human into a 3D Generative Model using
Pixel-aligned Reconstruction Priors [56.192682114114724]
Get3DHuman is a novel 3D human framework that can significantly boost the realism and diversity of the generated outcomes.
Our key observation is that the 3D generator can profit from human-related priors learned through 2D human generators and 3D reconstructors.
arXiv Detail & Related papers (2023-02-02T15:37:46Z) - Deep Generative Models on 3D Representations: A Survey [81.73385191402419]
Generative models aim to learn the distribution of observed data by generating new instances.
Recently, researchers started to shift focus from 2D to 3D space.
representing 3D data poses significantly greater challenges.
arXiv Detail & Related papers (2022-10-27T17:59:50Z) - EVA3D: Compositional 3D Human Generation from 2D Image Collections [27.70991135165909]
EVA3D is an unconditional 3D human generative model learned from 2D image collections only.
It can sample 3D humans with detailed geometry and render high-quality images (up to 512x256) without bells and whistles.
It achieves state-of-the-art 3D human generation performance regarding both geometry and texture quality.
arXiv Detail & Related papers (2022-10-10T17:59:31Z) - UltraPose: Synthesizing Dense Pose with 1 Billion Points by Human-body
Decoupling 3D Model [58.70130563417079]
We introduce a new 3D human-body model with a series of decoupled parameters that could freely control the generation of the body.
Compared to the existing manually annotated DensePose-COCO dataset, the synthetic UltraPose has ultra dense image-to-surface correspondences without annotation cost and error.
arXiv Detail & Related papers (2021-10-28T16:24:55Z) - Detailed 3D Human Body Reconstruction from Multi-view Images Combining
Voxel Super-Resolution and Learned Implicit Representation [12.459968574683625]
We propose a coarse-to-fine method to reconstruct a detailed 3D human body from multi-view images.
The coarse 3D models are estimated by learning an implicit representation based on multi-scale features.
The refined detailed 3D human body models can be produced by the voxel super-resolution which can preserve the details.
arXiv Detail & Related papers (2020-12-11T08:07:39Z) - Learning 3D Human Shape and Pose from Dense Body Parts [117.46290013548533]
We propose a Decompose-and-aggregate Network (DaNet) to learn 3D human shape and pose from dense correspondences of body parts.
Messages from local streams are aggregated to enhance the robust prediction of the rotation-based poses.
Our method is validated on both indoor and real-world datasets including Human3.6M, UP3D, COCO, and 3DPW.
arXiv Detail & Related papers (2019-12-31T15:09:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.