Self-Learning Transformations for Improving Gaze and Head Redirection
- URL: http://arxiv.org/abs/2010.12307v1
- Date: Fri, 23 Oct 2020 11:18:37 GMT
- Title: Self-Learning Transformations for Improving Gaze and Head Redirection
- Authors: Yufeng Zheng, Seonwook Park, Xucong Zhang, Shalini De Mello, Otmar
Hilliges
- Abstract summary: We propose a novel generative model for images of faces, that is capable of producing high-quality images under fine-grained control over eye gaze and head orientation angles.
This requires the disentangling of many appearance related factors including gaze and head orientation but also lighting, hue etc.
We show that explicitly disentangling task-irrelevant factors results in more accurate modelling of gaze and head orientation.
- Score: 49.61091281780071
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Many computer vision tasks rely on labeled data. Rapid progress in generative
modeling has led to the ability to synthesize photorealistic images. However,
controlling specific aspects of the generation process such that the data can
be used for supervision of downstream tasks remains challenging. In this paper
we propose a novel generative model for images of faces, that is capable of
producing high-quality images under fine-grained control over eye gaze and head
orientation angles. This requires the disentangling of many appearance related
factors including gaze and head orientation but also lighting, hue etc. We
propose a novel architecture which learns to discover, disentangle and encode
these extraneous variations in a self-learned manner. We further show that
explicitly disentangling task-irrelevant factors results in more accurate
modelling of gaze and head orientation. A novel evaluation scheme shows that
our method improves upon the state-of-the-art in redirection accuracy and
disentanglement between gaze direction and head orientation changes.
Furthermore, we show that in the presence of limited amounts of real-world
training data, our method allows for improvements in the downstream task of
semi-supervised cross-dataset gaze estimation. Please check our project page
at: https://ait.ethz.ch/projects/2020/STED-gaze/
Related papers
- EyeBAG: Accurate Control of Eye Blink and Gaze Based on Data
Augmentation Leveraging Style Mixing [0.483420384410068]
We introduce a novel framework consisting of two distinct modules: a blink control module and a gaze redirection module.
We show that our framework produces eye-controlled images of high quality, and demonstrate how it can be used to improve the performance of downstream tasks.
arXiv Detail & Related papers (2023-06-30T03:49:23Z) - Spatial Steerability of GANs via Self-Supervision from Discriminator [123.27117057804732]
We propose a self-supervised approach to improve the spatial steerability of GANs without searching for steerable directions in the latent space.
Specifically, we design randomly sampled Gaussian heatmaps to be encoded into the intermediate layers of generative models as spatial inductive bias.
During inference, users can interact with the spatial heatmaps in an intuitive manner, enabling them to edit the output image by adjusting the scene layout, moving, or removing objects.
arXiv Detail & Related papers (2023-01-20T07:36:29Z) - NeRF-Gaze: A Head-Eye Redirection Parametric Model for Gaze Estimation [37.977032771941715]
We propose a novel Head-Eye redirection parametric model based on Neural Radiance Field.
Our model can decouple the face and eyes for separate neural rendering.
It can achieve the purpose of separately controlling the attributes of the face, identity, illumination, and eye gaze direction.
arXiv Detail & Related papers (2022-12-30T13:52:28Z) - 3DGazeNet: Generalizing Gaze Estimation with Weak-Supervision from
Synthetic Views [67.00931529296788]
We propose to train general gaze estimation models which can be directly employed in novel environments without adaptation.
We create a large-scale dataset of diverse faces with gaze pseudo-annotations, which we extract based on the 3D geometry of the scene.
We test our method in the task of gaze generalization, in which we demonstrate improvement of up to 30% compared to state-of-the-art when no ground truth data are available.
arXiv Detail & Related papers (2022-12-06T14:15:17Z) - Masked Autoencoders are Robust Data Augmentors [90.34825840657774]
Regularization techniques like image augmentation are necessary for deep neural networks to generalize well.
We propose a novel perspective of augmentation to regularize the training process.
We show that utilizing such model-based nonlinear transformation as data augmentation can improve high-level recognition tasks.
arXiv Detail & Related papers (2022-06-10T02:41:48Z) - CUDA-GR: Controllable Unsupervised Domain Adaptation for Gaze
Redirection [3.0141238193080295]
The aim of gaze redirection is to manipulate the gaze in an image to the desired direction.
Advancement in generative adversarial networks has shown excellent results in generating photo-realistic images.
To enable such fine-tuned control, one needs to obtain ground truth annotations for the training data which can be very expensive.
arXiv Detail & Related papers (2021-06-21T04:39:42Z) - Unsupervised Discovery of Disentangled Manifolds in GANs [74.24771216154105]
Interpretable generation process is beneficial to various image editing applications.
We propose a framework to discover interpretable directions in the latent space given arbitrary pre-trained generative adversarial networks.
arXiv Detail & Related papers (2020-11-24T02:18:08Z) - 360-Degree Gaze Estimation in the Wild Using Multiple Zoom Scales [26.36068336169795]
We develop a model that mimics humans' ability to estimate the gaze by aggregating from focused looks.
The model avoids the need to extract clear eye patches.
We extend the model to handle the challenging task of 360-degree gaze estimation.
arXiv Detail & Related papers (2020-09-15T08:45:12Z) - Coarse-to-Fine Gaze Redirection with Numerical and Pictorial Guidance [74.27389895574422]
We propose a novel gaze redirection framework which exploits both a numerical and a pictorial direction guidance.
The proposed method outperforms the state-of-the-art approaches in terms of both image quality and redirection precision.
arXiv Detail & Related papers (2020-04-07T01:17:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.