Towards a Pipeline for Real-Time Visualization of Faces for VR-based
Telepresence and Live Broadcasting Utilizing Neural Rendering
- URL: http://arxiv.org/abs/2301.01490v1
- Date: Wed, 4 Jan 2023 08:49:51 GMT
- Title: Towards a Pipeline for Real-Time Visualization of Faces for VR-based
Telepresence and Live Broadcasting Utilizing Neural Rendering
- Authors: Philipp Ladwig, Rene Ebertowski, Alexander Pech, Ralf D\"orner,
Christian Geiger
- Abstract summary: Head-mounted displays (HMDs) for Virtual Reality pose a considerable obstacle for a realistic face-to-face conversation in VR.
We present an approach that focuses on low-cost hardware and can be used on a commodity gaming computer with a single GPU.
- Score: 58.720142291102135
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While head-mounted displays (HMDs) for Virtual Reality (VR) have become
widely available in the consumer market, they pose a considerable obstacle for
a realistic face-to-face conversation in VR since HMDs hide a significant
portion of the participants faces. Even with image streams from cameras
directly attached to an HMD, stitching together a convincing image of an entire
face remains a challenging task because of extreme capture angles and strong
lens distortions due to a wide field of view. Compared to the long line of
research in VR, reconstruction of faces hidden beneath an HMD is a very recent
topic of research. While the current state-of-the-art solutions demonstrate
photo-realistic 3D reconstruction results, they require high-cost laboratory
equipment and large computational costs. We present an approach that focuses on
low-cost hardware and can be used on a commodity gaming computer with a single
GPU. We leverage the benefits of an end-to-end pipeline by means of Generative
Adversarial Networks (GAN). Our GAN produces a frontal-facing 2.5D point cloud
based on a training dataset captured with an RGBD camera. In our approach, the
training process is offline, while the reconstruction runs in real-time. Our
results show adequate reconstruction quality within the 'learned' expressions.
Expressions not learned by the network produce artifacts and can trigger the
Uncanny Valley effect.
Related papers
- VOODOO XP: Expressive One-Shot Head Reenactment for VR Telepresence [14.010324388059866]
VOODOO XP is a 3D-aware one-shot head reenactment method that can generate highly expressive facial expressions from any input driver video and a single 2D portrait.
We show our solution on a monocular video setting and an end-to-end VR telepresence system for two-way communication.
arXiv Detail & Related papers (2024-05-25T12:33:40Z) - FNeVR: Neural Volume Rendering for Face Animation [53.92664037596834]
We propose a Face Neural Volume Rendering (FNeVR) network to explore the potential of 2D motion warping and 3D volume rendering.
In FNeVR, we design a 3D Face Volume Rendering (FVR) module to enhance the facial details for image rendering.
We also design a lightweight pose editor, enabling FNeVR to edit the facial pose in a simple yet effective way.
arXiv Detail & Related papers (2022-09-21T13:18:59Z) - NeuralPassthrough: Learned Real-Time View Synthesis for VR [3.907767419763815]
We propose the first learned passthrough method and assess its performance using a custom VR headset with a stereo pair of RGB cameras.
We demonstrate that our learned passthrough method delivers superior image quality compared to state-of-the-art methods.
arXiv Detail & Related papers (2022-07-05T17:39:22Z) - Robust Egocentric Photo-realistic Facial Expression Transfer for Virtual
Reality [68.18446501943585]
Social presence will fuel the next generation of communication systems driven by digital humans in virtual reality (VR)
The best 3D video-realistic VR avatars that minimize the uncanny effect rely on person-specific (PS) models.
This paper makes progress in overcoming these limitations by proposing an end-to-end multi-identity architecture.
arXiv Detail & Related papers (2021-04-10T15:48:53Z) - Pixel Codec Avatars [99.36561532588831]
Pixel Codec Avatars (PiCA) is a deep generative model of 3D human faces.
On a single Oculus Quest 2 mobile VR headset, 5 avatars are rendered in realtime in the same scene.
arXiv Detail & Related papers (2021-04-09T23:17:36Z) - High-fidelity Face Tracking for AR/VR via Deep Lighting Adaptation [117.32310997522394]
3D video avatars can empower virtual communications by providing compression, privacy, entertainment, and a sense of presence in AR/VR.
Existing person-specific 3D models are not robust to lighting, hence their results typically miss subtle facial behaviors and cause artifacts in the avatar.
This paper addresses previous limitations by learning a deep learning lighting model, that in combination with a high-quality 3D face tracking algorithm, provides a method for subtle and robust facial motion transfer from a regular video to a 3D photo-realistic avatar.
arXiv Detail & Related papers (2021-03-29T18:33:49Z) - Unmasking Communication Partners: A Low-Cost AI Solution for Digitally
Removing Head-Mounted Displays in VR-Based Telepresence [62.997667081978825]
Face-to-face conversation in Virtual Reality (VR) is a challenge when participants wear head-mounted displays (HMD)
Past research has shown that high-fidelity face reconstruction with personal avatars in VR is possible under laboratory conditions with high-cost hardware.
We propose one of the first low-cost systems for this task which uses only open source, free software and affordable hardware.
arXiv Detail & Related papers (2020-11-06T23:17:12Z) - Facial Expression Recognition Under Partial Occlusion from Virtual
Reality Headsets based on Transfer Learning [0.0]
convolutional neural network based approaches has become widely adopted due to their proven applicability to Facial Expression Recognition task.
However, recognizing facial expression while wearing a head-mounted VR headset is a challenging task due to the upper half of the face being completely occluded.
We propose a geometric model to simulate occlusion resulting from a Samsung Gear VR headset that can be applied to existing FER datasets.
arXiv Detail & Related papers (2020-08-12T20:25:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.