EgoRenderer: Rendering Human Avatars from Egocentric Camera Images
- URL: http://arxiv.org/abs/2111.12685v1
- Date: Wed, 24 Nov 2021 18:33:02 GMT
- Title: EgoRenderer: Rendering Human Avatars from Egocentric Camera Images
- Authors: Tao Hu, Kripasindhu Sarkar, Lingjie Liu, Matthias Zwicker, Christian
Theobalt
- Abstract summary: We present EgoRenderer, a system for rendering full-body neural avatars of a person captured by a wearable, egocentric fisheye camera.
Rendering full-body avatars from such egocentric images come with unique challenges due to the top-down view and large distortions.
We tackle these challenges by decomposing the rendering process into several steps, including texture synthesis, pose construction, and neural image translation.
- Score: 87.96474006263692
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present EgoRenderer, a system for rendering full-body neural avatars of a
person captured by a wearable, egocentric fisheye camera that is mounted on a
cap or a VR headset. Our system renders photorealistic novel views of the actor
and her motion from arbitrary virtual camera locations. Rendering full-body
avatars from such egocentric images come with unique challenges due to the
top-down view and large distortions. We tackle these challenges by decomposing
the rendering process into several steps, including texture synthesis, pose
construction, and neural image translation. For texture synthesis, we propose
Ego-DPNet, a neural network that infers dense correspondences between the input
fisheye images and an underlying parametric body model, and to extract textures
from egocentric inputs. In addition, to encode dynamic appearances, our
approach also learns an implicit texture stack that captures detailed
appearance variation across poses and viewpoints. For correct pose generation,
we first estimate body pose from the egocentric view using a parametric model.
We then synthesize an external free-viewpoint pose image by projecting the
parametric model to the user-specified target viewpoint. We next combine the
target pose image and the textures into a combined feature image, which is
transformed into the output color image using a neural image translation
network. Experimental evaluations show that EgoRenderer is capable of
generating realistic free-viewpoint avatars of a person wearing an egocentric
camera. Comparisons to several baselines demonstrate the advantages of our
approach.
Related papers
- EgoAvatar: Egocentric View-Driven and Photorealistic Full-body Avatars [56.56236652774294]
We propose a person-specific egocentric telepresence approach, which jointly models the photoreal digital avatar while also driving it from a single egocentric video.
Our experiments demonstrate a clear step towards egocentric and photoreal telepresence as our method outperforms baselines as well as competing methods.
arXiv Detail & Related papers (2024-09-22T22:50:27Z) - TexVocab: Texture Vocabulary-conditioned Human Avatars [42.170169762733835]
TexVocab is a novel avatar representation that constructs a texture vocabulary and associates body poses with texture maps for animation.
Our method is able to create animatable human avatars with detailed and dynamic appearances from RGB videos.
arXiv Detail & Related papers (2024-03-31T01:58:04Z) - FLARE: Fast Learning of Animatable and Relightable Mesh Avatars [64.48254296523977]
Our goal is to efficiently learn personalized animatable 3D head avatars from videos that are geometrically accurate, realistic, relightable, and compatible with current rendering systems.
We introduce FLARE, a technique that enables the creation of animatable and relightable avatars from a single monocular video.
arXiv Detail & Related papers (2023-10-26T16:13:00Z) - Scene-aware Egocentric 3D Human Pose Estimation [72.57527706631964]
Egocentric 3D human pose estimation with a single head-mounted fisheye camera has recently attracted attention due to its numerous applications in virtual and augmented reality.
Existing methods still struggle in challenging poses where the human body is highly occluded or is closely interacting with the scene.
We propose a scene-aware egocentric pose estimation method that guides the prediction of the egocentric pose with scene constraints.
arXiv Detail & Related papers (2022-12-20T21:35:39Z) - Realistic One-shot Mesh-based Head Avatars [7.100064936484693]
We present a system for realistic one-shot mesh-based human head avatars creation, ROME for short.
Using a single photograph, our model estimates a person-specific head mesh and the associated neural texture, which encodes both local photometric and geometric details.
The resulting avatars are rigged and can be rendered using a neural network, which is trained alongside the mesh and texture estimators on a dataset of in-the-wild videos.
arXiv Detail & Related papers (2022-06-16T17:45:23Z) - Pipeline for 3D reconstruction of the human body from AR/VR headset
mounted egocentric cameras [0.0]
We propose a novel pipeline for the 3D reconstruction of the full body from egocentric viewpoints.
We first make use of conditional GANs to translate the egocentric views to full body third-person views.
The generated mesh has fairly realistic body proportions and is fully rigged allowing for further applications.
arXiv Detail & Related papers (2021-11-09T20:38:32Z) - Neural Re-Rendering of Humans from a Single Image [80.53438609047896]
We propose a new method for neural re-rendering of a human under a novel user-defined pose and viewpoint.
Our algorithm represents body pose and shape as a parametric mesh which can be reconstructed from a single image.
arXiv Detail & Related papers (2021-01-11T18:53:47Z) - SelfPose: 3D Egocentric Pose Estimation from a Headset Mounted Camera [97.0162841635425]
We present a solution to egocentric 3D body pose estimation from monocular images captured from downward looking fish-eye cameras installed on the rim of a head mounted VR device.
This unusual viewpoint leads to images with unique visual appearance, with severe self-occlusions and perspective distortions.
We propose an encoder-decoder architecture with a novel multi-branch decoder designed to account for the varying uncertainty in 2D predictions.
arXiv Detail & Related papers (2020-11-02T16:18:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.