MonoNHR: Monocular Neural Human Renderer
- URL: http://arxiv.org/abs/2210.00627v1
- Date: Sun, 2 Oct 2022 21:01:02 GMT
- Title: MonoNHR: Monocular Neural Human Renderer
- Authors: Hongsuk Choi, Gyeongsik Moon, Matthieu Armando, Vincent Leroy, Kyoung
Mu Lee, Gregory Rogez
- Abstract summary: We propose Monocular Neural Human Renderer (MonoNHR), a novel approach that renders robust free-viewpoint images of an arbitrary human given only a single image.
First, we propose to disentangle 3D geometry and texture features and to condition the texture inference on the 3D geometry features.
Second, we introduce a Mesh Inpainter module that inpaints the occluded parts exploiting human structural priors such as symmetry.
- Score: 51.396845817689915
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing neural human rendering methods struggle with a single image input
due to the lack of information in invisible areas and the depth ambiguity of
pixels in visible areas. In this regard, we propose Monocular Neural Human
Renderer (MonoNHR), a novel approach that renders robust free-viewpoint images
of an arbitrary human given only a single image. MonoNHR is the first method
that (i) renders human subjects never seen during training in a monocular
setup, and (ii) is trained in a weakly-supervised manner without geometry
supervision. First, we propose to disentangle 3D geometry and texture features
and to condition the texture inference on the 3D geometry features. Second, we
introduce a Mesh Inpainter module that inpaints the occluded parts exploiting
human structural priors such as symmetry. Experiments on ZJU-MoCap, AIST, and
HUMBI datasets show that our approach significantly outperforms the recent
methods adapted to the monocular case.
Related papers
- Single Image, Any Face: Generalisable 3D Face Generation [59.9369171926757]
We propose a novel model, Gen3D-Face, which generates 3D human faces with unconstrained single image input.
To the best of our knowledge, this is the first attempt and benchmark for creating photorealistic 3D human face avatars from single images.
arXiv Detail & Related papers (2024-09-25T14:56:37Z) - COSMU: Complete 3D human shape from monocular unconstrained images [24.08612483445495]
We present a novel framework to reconstruct complete 3D human shapes from a given target image by leveraging monocular unconstrained images.
The objective of this work is to reproduce high-quality details in regions of the reconstructed human body that are not visible in the input target.
arXiv Detail & Related papers (2024-07-15T10:06:59Z) - HINT: Learning Complete Human Neural Representations from Limited Viewpoints [69.76947323932107]
We propose a NeRF-based algorithm able to learn a detailed and complete human model from limited viewing angles.
As a result, our method can reconstruct complete humans even from a few viewing angles, increasing performance by more than 15% PSNR.
arXiv Detail & Related papers (2024-05-30T05:43:09Z) - HR Human: Modeling Human Avatars with Triangular Mesh and High-Resolution Textures from Videos [52.23323966700072]
We present a framework for acquiring human avatars that are attached with high-resolution physically-based material textures and mesh from monocular video.
Our method introduces a novel information fusion strategy to combine the information from the monocular video and synthesize virtual multi-view images.
Experiments show that our approach outperforms previous representations in terms of high fidelity, and this explicit result supports deployment on common triangulars.
arXiv Detail & Related papers (2024-05-18T11:49:09Z) - SiTH: Single-view Textured Human Reconstruction with Image-Conditioned Diffusion [35.73448283467723]
SiTH is a novel pipeline that integrates an image-conditioned diffusion model into a 3D mesh reconstruction workflow.
We employ a powerful generative diffusion model to hallucinate unseen back-view appearance based on the input images.
For the latter, we leverage skinned body meshes as guidance to recover full-body texture meshes from the input and back-view images.
arXiv Detail & Related papers (2023-11-27T14:22:07Z) - Rendering Humans from Object-Occluded Monocular Videos [32.67336188239284]
3D understanding and rendering of moving humans from monocular videos is a challenging task.
Existing methods cannot handle such defects due to two reasons.
We present OccNeRF, a neural rendering method that achieves better rendering of humans in severely occluded scenes.
arXiv Detail & Related papers (2023-08-08T23:12:33Z) - One-shot Implicit Animatable Avatars with Model-based Priors [31.385051428938585]
ELICIT is a novel method for learning human-specific neural radiance fields from a single image.
ELICIT has outperformed strong baseline methods of avatar creation when only a single image is available.
arXiv Detail & Related papers (2022-12-05T18:24:06Z) - MagicPony: Learning Articulated 3D Animals in the Wild [81.63322697335228]
We present a new method, dubbed MagicPony, that learns this predictor purely from in-the-wild single-view images of the object category.
At its core is an implicit-explicit representation of articulated shape and appearance, combining the strengths of neural fields and meshes.
arXiv Detail & Related papers (2022-11-22T18:59:31Z) - DRaCoN -- Differentiable Rasterization Conditioned Neural Radiance
Fields for Articulated Avatars [92.37436369781692]
We present DRaCoN, a framework for learning full-body volumetric avatars.
It exploits the advantages of both the 2D and 3D neural rendering techniques.
Experiments on the challenging ZJU-MoCap and Human3.6M datasets indicate that DRaCoN outperforms state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T17:59:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.