High-fidelity Face Tracking for AR/VR via Deep Lighting Adaptation
- URL: http://arxiv.org/abs/2103.15876v1
- Date: Mon, 29 Mar 2021 18:33:49 GMT
- Title: High-fidelity Face Tracking for AR/VR via Deep Lighting Adaptation
- Authors: Lele Chen, Chen Cao, Fernando De la Torre, Jason Saragih, Chenliang
Xu, Yaser Sheikh
- Abstract summary: 3D video avatars can empower virtual communications by providing compression, privacy, entertainment, and a sense of presence in AR/VR.
Existing person-specific 3D models are not robust to lighting, hence their results typically miss subtle facial behaviors and cause artifacts in the avatar.
This paper addresses previous limitations by learning a deep learning lighting model, that in combination with a high-quality 3D face tracking algorithm, provides a method for subtle and robust facial motion transfer from a regular video to a 3D photo-realistic avatar.
- Score: 117.32310997522394
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: 3D video avatars can empower virtual communications by providing compression,
privacy, entertainment, and a sense of presence in AR/VR. Best 3D
photo-realistic AR/VR avatars driven by video, that can minimize uncanny
effects, rely on person-specific models. However, existing person-specific
photo-realistic 3D models are not robust to lighting, hence their results
typically miss subtle facial behaviors and cause artifacts in the avatar. This
is a major drawback for the scalability of these models in communication
systems (e.g., Messenger, Skype, FaceTime) and AR/VR. This paper addresses
previous limitations by learning a deep learning lighting model, that in
combination with a high-quality 3D face tracking algorithm, provides a method
for subtle and robust facial motion transfer from a regular video to a 3D
photo-realistic avatar. Extensive experimental validation and comparisons to
other state-of-the-art methods demonstrate the effectiveness of the proposed
framework in real-world scenarios with variability in pose, expression, and
illumination. Please visit https://www.youtube.com/watch?v=dtz1LgZR8cc for more
results. Our project page can be found at
https://www.cs.rochester.edu/u/lchen63.
Related papers
- UltrAvatar: A Realistic Animatable 3D Avatar Diffusion Model with Authenticity Guided Textures [80.047065473698]
We propose a novel 3D avatar generation approach termed UltrAvatar with enhanced fidelity of geometry, and superior quality of physically based rendering (PBR) textures without unwanted lighting.
We demonstrate the effectiveness and robustness of the proposed method, outperforming the state-of-the-art methods by a large margin in the experiments.
arXiv Detail & Related papers (2024-01-20T01:55:17Z) - Reality's Canvas, Language's Brush: Crafting 3D Avatars from Monocular Video [14.140380599168628]
ReCaLaB is a pipeline that learns high-fidelity 3D human avatars from just a single RGB video.
A pose-conditioned NeRF is optimized to volumetrically represent a human subject in canonical T-pose.
An image-conditioned diffusion model thereby helps to animate appearance and pose of the 3D avatar to create video sequences with previously unseen human motion.
arXiv Detail & Related papers (2023-12-08T01:53:06Z) - GAN-Avatar: Controllable Personalized GAN-based Human Head Avatar [48.21353924040671]
We propose to learn person-specific animatable avatars from images without assuming to have access to precise facial expression tracking.
We learn a mapping from 3DMM facial expression parameters to the latent space of the generative model.
With this scheme, we decouple 3D appearance reconstruction and animation control to achieve high fidelity in image synthesis.
arXiv Detail & Related papers (2023-11-22T19:13:00Z) - DRaCoN -- Differentiable Rasterization Conditioned Neural Radiance
Fields for Articulated Avatars [92.37436369781692]
We present DRaCoN, a framework for learning full-body volumetric avatars.
It exploits the advantages of both the 2D and 3D neural rendering techniques.
Experiments on the challenging ZJU-MoCap and Human3.6M datasets indicate that DRaCoN outperforms state-of-the-art methods.
arXiv Detail & Related papers (2022-03-29T17:59:15Z) - Robust Egocentric Photo-realistic Facial Expression Transfer for Virtual
Reality [68.18446501943585]
Social presence will fuel the next generation of communication systems driven by digital humans in virtual reality (VR)
The best 3D video-realistic VR avatars that minimize the uncanny effect rely on person-specific (PS) models.
This paper makes progress in overcoming these limitations by proposing an end-to-end multi-identity architecture.
arXiv Detail & Related papers (2021-04-10T15:48:53Z) - Pixel Codec Avatars [99.36561532588831]
Pixel Codec Avatars (PiCA) is a deep generative model of 3D human faces.
On a single Oculus Quest 2 mobile VR headset, 5 avatars are rendered in realtime in the same scene.
arXiv Detail & Related papers (2021-04-09T23:17:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.