PeeledHuman: Robust Shape Representation for Textured 3D Human Body
Reconstruction
- URL: http://arxiv.org/abs/2002.06664v2
- Date: Mon, 2 Nov 2020 10:06:42 GMT
- Title: PeeledHuman: Robust Shape Representation for Textured 3D Human Body
Reconstruction
- Authors: Sai Sagar Jinka, Rohan Chacko, Avinash Sharma and P. J. Narayanan
- Abstract summary: PeeledHuman encodes the human body as a set of Peeled Depth and RGB maps in 2D.
We train PeelGAN using a 3D Chamfer loss and other 2D losses to generate multiple depth values per-pixel and a corresponding RGB field per-vertex.
In our simple non-parametric solution, the generated Peeled Depth maps are back-projected to 3D space to obtain a complete textured 3D shape.
- Score: 7.582064461041252
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce PeeledHuman - a novel shape representation of the human body
that is robust to self-occlusions. PeeledHuman encodes the human body as a set
of Peeled Depth and RGB maps in 2D, obtained by performing ray-tracing on the
3D body model and extending each ray beyond its first intersection. This
formulation allows us to handle self-occlusions efficiently compared to other
representations. Given a monocular RGB image, we learn these Peeled maps in an
end-to-end generative adversarial fashion using our novel framework - PeelGAN.
We train PeelGAN using a 3D Chamfer loss and other 2D losses to generate
multiple depth values per-pixel and a corresponding RGB field per-vertex in a
dual-branch setup. In our simple non-parametric solution, the generated Peeled
Depth maps are back-projected to 3D space to obtain a complete textured 3D
shape. The corresponding RGB maps provide vertex-level texture details. We
compare our method with current parametric and non-parametric methods in 3D
reconstruction and find that we achieve state-of-the-art-results. We
demonstrate the effectiveness of our representation on publicly available BUFF
and MonoPerfCap datasets as well as loose clothing data collected by our
calibrated multi-Kinect setup.
Related papers
- FAMOUS: High-Fidelity Monocular 3D Human Digitization Using View Synthesis [51.193297565630886]
The challenge of accurately inferring texture remains, particularly in obscured areas such as the back of a person in frontal-view images.
This limitation in texture prediction largely stems from the scarcity of large-scale and diverse 3D datasets.
We propose leveraging extensive 2D fashion datasets to enhance both texture and shape prediction in 3D human digitization.
arXiv Detail & Related papers (2024-10-13T01:25:05Z) - Sampling is Matter: Point-guided 3D Human Mesh Reconstruction [0.0]
This paper presents a simple yet powerful method for 3D human mesh reconstruction from a single RGB image.
Experimental results on benchmark datasets show that the proposed method efficiently improves the performance of 3D human mesh reconstruction.
arXiv Detail & Related papers (2023-04-19T08:45:26Z) - SSR-2D: Semantic 3D Scene Reconstruction from 2D Images [54.46126685716471]
In this work, we explore a central 3D scene modeling task, namely, semantic scene reconstruction without using any 3D annotations.
The key idea of our approach is to design a trainable model that employs both incomplete 3D reconstructions and their corresponding source RGB-D images.
Our method achieves the state-of-the-art performance of semantic scene completion on two large-scale benchmark datasets MatterPort3D and ScanNet.
arXiv Detail & Related papers (2023-02-07T17:47:52Z) - Multi-Person 3D Pose and Shape Estimation via Inverse Kinematics and
Refinement [5.655207244072081]
Estimating 3D poses and shapes in the form of meshes from monocular RGB images is challenging.
We propose a coarse-to-fine pipeline that benefits from 1) inverse kinematics from the occlusion-robust 3D skeleton estimation.
We demonstrate the effectiveness of our method, outperforming state-of-the-arts on 3DPW, MuPoTS and AGORA datasets.
arXiv Detail & Related papers (2022-10-24T18:29:06Z) - MvDeCor: Multi-view Dense Correspondence Learning for Fine-grained 3D
Segmentation [91.6658845016214]
We propose to utilize self-supervised techniques in the 2D domain for fine-grained 3D shape segmentation tasks.
We render a 3D shape from multiple views, and set up a dense correspondence learning task within the contrastive learning framework.
As a result, the learned 2D representations are view-invariant and geometrically consistent.
arXiv Detail & Related papers (2022-08-18T00:48:15Z) - SHARP: Shape-Aware Reconstruction of People in Loose Clothing [6.469298908778292]
SHARP (SHape Aware Reconstruction of People in loose clothing) is a novel end-to-end trainable network.
It recovers the 3D geometry and appearance of humans in loose clothing from a monocular image.
We show superior qualitative and quantitative performance than existing state-of-the-art methods.
arXiv Detail & Related papers (2022-05-24T10:26:42Z) - DRaCoN -- Differentiable Rasterization Conditioned Neural Radiance
Fields for Articulated Avatars [92.37436369781692]
We present DRaCoN, a framework for learning full-body volumetric avatars.
It exploits the advantages of both the 2D and 3D neural rendering techniques.
Experiments on the challenging ZJU-MoCap and Human3.6M datasets indicate that DRaCoN outperforms state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T17:59:15Z) - Tracking People with 3D Representations [78.97070307547283]
We present a novel approach for tracking multiple people in video.
Unlike past approaches which employ 2D representations, we employ 3D representations of people, located in three-dimensional space.
We find that 3D representations are more effective than 2D representations for tracking in these settings.
arXiv Detail & Related papers (2021-11-15T16:15:21Z) - SHARP: Shape-Aware Reconstruction of People In Loose Clothing [6.796748304066826]
3D human body reconstruction from monocular images is an interesting and ill-posed problem in computer vision.
We propose SHARP, a novel end-to-end trainable network that accurately recovers the detailed geometry and appearance of 3D people in loose clothing from a monocular image.
We evaluate SHARP on publicly available Cloth3D and THuman datasets and report superior performance to state-of-the-art approaches.
arXiv Detail & Related papers (2021-06-09T02:54:53Z) - Synthetic Training for Monocular Human Mesh Recovery [100.38109761268639]
This paper aims to estimate 3D mesh of multiple body parts with large-scale differences from a single RGB image.
The main challenge is lacking training data that have complete 3D annotations of all body parts in 2D images.
We propose a depth-to-scale (D2S) projection to incorporate the depth difference into the projection function to derive per-joint scale variants.
arXiv Detail & Related papers (2020-10-27T03:31:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.