Attention based Occlusion Removal for Hybrid Telepresence Systems
- URL: http://arxiv.org/abs/2112.01098v1
- Date: Thu, 2 Dec 2021 10:18:22 GMT
- Title: Attention based Occlusion Removal for Hybrid Telepresence Systems
- Authors: Surabhi Gupta, Ashwath Shetty, Avinash Sharma
- Abstract summary: We propose a novel attention-enabled encoder-decoder architecture for HMD de-occlusion.
We report superior qualitative and quantitative results over state-of-the-art methods.
We also present applications of this approach to hybrid video teleconferencing using existing animation and 3D face reconstruction pipelines.
- Score: 5.006086647446482
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Traditionally, video conferencing is a widely adopted solution for
telecommunication, but a lack of immersiveness comes inherently due to the 2D
nature of facial representation. The integration of Virtual Reality (VR) in a
communication/telepresence system through Head Mounted Displays (HMDs) promises
to provide users a much better immersive experience. However, HMDs cause
hindrance by blocking the facial appearance and expressions of the user. To
overcome these issues, we propose a novel attention-enabled encoder-decoder
architecture for HMD de-occlusion. We also propose to train our person-specific
model using short videos (1-2 minutes) of the user, captured in varying
appearances, and demonstrated generalization to unseen poses and appearances of
the user. We report superior qualitative and quantitative results over
state-of-the-art methods. We also present applications of this approach to
hybrid video teleconferencing using existing animation and 3D face
reconstruction pipelines.
Related papers
- DICE: End-to-end Deformation Capture of Hand-Face Interactions from a Single Image [98.29284902879652]
We present DICE, the first end-to-end method for Deformation-aware hand-face Interaction reCovEry from a single image.
It features disentangling the regression of local deformation fields and global mesh locations into two network branches.
It achieves state-of-the-art performance on a standard benchmark and in-the-wild data in terms of accuracy and physical plausibility.
arXiv Detail & Related papers (2024-06-26T00:08:29Z) - VOODOO XP: Expressive One-Shot Head Reenactment for VR Telepresence [14.010324388059866]
VOODOO XP is a 3D-aware one-shot head reenactment method that can generate highly expressive facial expressions from any input driver video and a single 2D portrait.
We show our solution on a monocular video setting and an end-to-end VR telepresence system for two-way communication.
arXiv Detail & Related papers (2024-05-25T12:33:40Z) - Expression-aware video inpainting for HMD removal in XR applications [0.27624021966289597]
Head-mounted displays (HMDs) serve as indispensable devices for observing extended reality (XR) environments and virtual content.
HMDs present an obstacle to external recording techniques as they block the upper face of the user.
We propose a new network for expression-aware video inpainting for HMD removal based on generative adversarial networks (GANs)
arXiv Detail & Related papers (2024-01-25T12:32:21Z) - 3D Human Pose Perception from Egocentric Stereo Videos [67.9563319914377]
We propose a new transformer-based framework to improve egocentric stereo 3D human pose estimation.
Our method is able to accurately estimate human poses even in challenging scenarios, such as crouching and sitting.
We will release UnrealEgo2, UnrealEgo-RW, and trained models on our project page.
arXiv Detail & Related papers (2023-12-30T21:21:54Z) - HMP: Hand Motion Priors for Pose and Shape Estimation from Video [52.39020275278984]
We develop a generative motion prior specific for hands, trained on the AMASS dataset which features diverse and high-quality hand motions.
Our integration of a robust motion prior significantly enhances performance, especially in occluded scenarios.
We demonstrate our method's efficacy via qualitative and quantitative evaluations on the HO3D and DexYCB datasets.
arXiv Detail & Related papers (2023-12-27T22:35:33Z) - Towards a Pipeline for Real-Time Visualization of Faces for VR-based
Telepresence and Live Broadcasting Utilizing Neural Rendering [58.720142291102135]
Head-mounted displays (HMDs) for Virtual Reality pose a considerable obstacle for a realistic face-to-face conversation in VR.
We present an approach that focuses on low-cost hardware and can be used on a commodity gaming computer with a single GPU.
arXiv Detail & Related papers (2023-01-04T08:49:51Z) - Pixel Codec Avatars [99.36561532588831]
Pixel Codec Avatars (PiCA) is a deep generative model of 3D human faces.
On a single Oculus Quest 2 mobile VR headset, 5 avatars are rendered in realtime in the same scene.
arXiv Detail & Related papers (2021-04-09T23:17:36Z) - Unmasking Communication Partners: A Low-Cost AI Solution for Digitally
Removing Head-Mounted Displays in VR-Based Telepresence [62.997667081978825]
Face-to-face conversation in Virtual Reality (VR) is a challenge when participants wear head-mounted displays (HMD)
Past research has shown that high-fidelity face reconstruction with personal avatars in VR is possible under laboratory conditions with high-cost hardware.
We propose one of the first low-cost systems for this task which uses only open source, free software and affordable hardware.
arXiv Detail & Related papers (2020-11-06T23:17:12Z) - Facial Expression Recognition Under Partial Occlusion from Virtual
Reality Headsets based on Transfer Learning [0.0]
convolutional neural network based approaches has become widely adopted due to their proven applicability to Facial Expression Recognition task.
However, recognizing facial expression while wearing a head-mounted VR headset is a challenging task due to the upper half of the face being completely occluded.
We propose a geometric model to simulate occlusion resulting from a Samsung Gear VR headset that can be applied to existing FER datasets.
arXiv Detail & Related papers (2020-08-12T20:25:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.