HQ3DAvatar: High Quality Controllable 3D Head Avatar
- URL: http://arxiv.org/abs/2303.14471v1
- Date: Sat, 25 Mar 2023 13:56:33 GMT
- Title: HQ3DAvatar: High Quality Controllable 3D Head Avatar
- Authors: Kartik Teotia, Mallikarjun B R, Xingang Pan, Hyeongwoo Kim, Pablo
Garrido, Mohamed Elgharib, Christian Theobalt
- Abstract summary: This paper presents a novel approach to building highly photorealistic digital head avatars.
Our method learns a canonical space via an implicit function parameterized by a neural network.
At test time, our method is driven by a monocular RGB video.
- Score: 65.70885416855782
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Multi-view volumetric rendering techniques have recently shown great
potential in modeling and synthesizing high-quality head avatars. A common
approach to capture full head dynamic performances is to track the underlying
geometry using a mesh-based template or 3D cube-based graphics primitives.
While these model-based approaches achieve promising results, they often fail
to learn complex geometric details such as the mouth interior, hair, and
topological changes over time. This paper presents a novel approach to building
highly photorealistic digital head avatars. Our method learns a canonical space
via an implicit function parameterized by a neural network. It leverages
multiresolution hash encoding in the learned feature space, allowing for
high-quality, faster training and high-resolution rendering. At test time, our
method is driven by a monocular RGB video. Here, an image encoder extracts
face-specific features that also condition the learnable canonical space. This
encourages deformation-dependent texture variations during training. We also
propose a novel optical flow based loss that ensures correspondences in the
learned canonical space, thus encouraging artifact-free and temporally
consistent renderings. We show results on challenging facial expressions and
show free-viewpoint renderings at interactive real-time rates for medium image
resolutions. Our method outperforms all existing approaches, both visually and
numerically. We will release our multiple-identity dataset to encourage further
research. Our Project page is available at:
https://vcai.mpi-inf.mpg.de/projects/HQ3DAvatar/
Related papers
- Efficient 3D Implicit Head Avatar with Mesh-anchored Hash Table Blendshapes [40.35875246929843]
3D head avatars built with neural implicit representations have achieved unprecedented levels of photorealism.
We propose a novel fast 3D neural implicit head avatar model that achieves real-time rendering while maintaining fine-grained controllability and high rendering quality.
arXiv Detail & Related papers (2024-04-02T00:55:50Z) - Bridging 3D Gaussian and Mesh for Freeview Video Rendering [57.21847030980905]
GauMesh bridges the 3D Gaussian and Mesh for modeling and rendering the dynamic scenes.
We show that our approach adapts the appropriate type of primitives to represent the different parts of the dynamic scene.
arXiv Detail & Related papers (2024-03-18T04:01:26Z) - FLARE: Fast Learning of Animatable and Relightable Mesh Avatars [64.48254296523977]
Our goal is to efficiently learn personalized animatable 3D head avatars from videos that are geometrically accurate, realistic, relightable, and compatible with current rendering systems.
We introduce FLARE, a technique that enables the creation of animatable and relightable avatars from a single monocular video.
arXiv Detail & Related papers (2023-10-26T16:13:00Z) - PonderV2: Pave the Way for 3D Foundation Model with A Universal
Pre-training Paradigm [114.47216525866435]
We introduce a novel universal 3D pre-training framework designed to facilitate the acquisition of efficient 3D representation.
For the first time, PonderV2 achieves state-of-the-art performance on 11 indoor and outdoor benchmarks, implying its effectiveness.
arXiv Detail & Related papers (2023-10-12T17:59:57Z) - Learning Personalized High Quality Volumetric Head Avatars from
Monocular RGB Videos [47.94545609011594]
We propose a method to learn a high-quality implicit 3D head avatar from a monocular RGB video captured in the wild.
Our hybrid pipeline combines the geometry prior and dynamic tracking of a 3DMM with a neural radiance field to achieve fine-grained control and photorealism.
arXiv Detail & Related papers (2023-04-04T01:10:04Z) - Human Performance Modeling and Rendering via Neural Animated Mesh [40.25449482006199]
We bridge the traditional mesh with a new class of neural rendering.
In this paper, we present a novel approach for rendering human views from video.
We demonstrate our approach on various platforms, inserting virtual human performances into AR headsets.
arXiv Detail & Related papers (2022-09-18T03:58:00Z) - Vision Transformer for NeRF-Based View Synthesis from a Single Input
Image [49.956005709863355]
We propose to leverage both the global and local features to form an expressive 3D representation.
To synthesize a novel view, we train a multilayer perceptron (MLP) network conditioned on the learned 3D representation to perform volume rendering.
Our method can render novel views from only a single input image and generalize across multiple object categories using a single model.
arXiv Detail & Related papers (2022-07-12T17:52:04Z) - KeypointNeRF: Generalizing Image-based Volumetric Avatars using Relative
Spatial Encoding of Keypoints [28.234772596912165]
We propose a highly effective approach to modeling high-fidelity volumetric avatars from sparse views.
One of the key ideas is to encode relative spatial 3D information via sparse 3D keypoints.
Our experiments show that a majority of errors in prior work stem from an inappropriate choice of spatial encoding.
arXiv Detail & Related papers (2022-05-10T15:57:03Z) - DRaCoN -- Differentiable Rasterization Conditioned Neural Radiance
Fields for Articulated Avatars [92.37436369781692]
We present DRaCoN, a framework for learning full-body volumetric avatars.
It exploits the advantages of both the 2D and 3D neural rendering techniques.
Experiments on the challenging ZJU-MoCap and Human3.6M datasets indicate that DRaCoN outperforms state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T17:59:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.