LEGAN: Disentangled Manipulation of Directional Lighting and Facial
Expressions by Leveraging Human Perceptual Judgements
- URL: http://arxiv.org/abs/2010.01464v3
- Date: Fri, 18 Jun 2021 22:00:59 GMT
- Title: LEGAN: Disentangled Manipulation of Directional Lighting and Facial
Expressions by Leveraging Human Perceptual Judgements
- Authors: Sandipan Banerjee, Ajjen Joshi, Prashant Mahajan, Sneha Bhattacharya,
Survi Kyal and Taniya Mishra
- Abstract summary: We propose LEGAN, a novel synthesis framework that leverages perceptual quality judgments for jointly manipulating lighting and expressions in face images.
LEGAN disentangles the lighting and expression subspaces and performs transformations in the feature space before upscaling to the desired output image.
We also conduct a perceptual study using images synthesized by LEGAN and other GAN models and show the correlation between our quality estimation and visual fidelity.
- Score: 7.5603864775031004
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Building facial analysis systems that generalize to extreme variations in
lighting and facial expressions is a challenging problem that can potentially
be alleviated using natural-looking synthetic data. Towards that, we propose
LEGAN, a novel synthesis framework that leverages perceptual quality judgments
for jointly manipulating lighting and expressions in face images, without
requiring paired training data. LEGAN disentangles the lighting and expression
subspaces and performs transformations in the feature space before upscaling to
the desired output image. The fidelity of the synthetic image is further
refined by integrating a perceptual quality estimation model, trained with face
images rendered using multiple synthesis methods and their crowd-sourced
naturalness ratings, into the LEGAN framework as an auxiliary discriminator.
Using objective metrics like FID and LPIPS, LEGAN is shown to generate higher
quality face images when compared with popular GAN models like StarGAN and
StarGAN-v2 for lighting and expression synthesis. We also conduct a perceptual
study using images synthesized by LEGAN and other GAN models and show the
correlation between our quality estimation and visual fidelity. Finally, we
demonstrate the effectiveness of LEGAN as training data augmenter for
expression recognition and face verification tasks.
Related papers
- SynFace: Face Recognition with Synthetic Data [83.15838126703719]
We devise the SynFace with identity mixup (IM) and domain mixup (DM) to mitigate the performance gap.
We also perform a systematically empirical analysis on synthetic face images to provide some insights on how to effectively utilize synthetic data for face recognition.
arXiv Detail & Related papers (2021-08-18T03:41:54Z) - Simultaneous Face Hallucination and Translation for Thermal to Visible
Face Verification using Axial-GAN [74.22129648654783]
We introduce the task of thermal-to-visible face verification from low-resolution thermal images.
We propose Axial-Generative Adversarial Network (Axial-GAN) to synthesize high-resolution visible images for matching.
arXiv Detail & Related papers (2021-04-13T22:34:28Z) - IMAGINE: Image Synthesis by Image-Guided Model Inversion [79.4691654458141]
We introduce an inversion based method, denoted as IMAge-Guided model INvErsion (IMAGINE), to generate high-quality and diverse images.
We leverage the knowledge of image semantics from a pre-trained classifier to achieve plausible generations.
IMAGINE enables the synthesis procedure to simultaneously 1) enforce semantic specificity constraints during the synthesis, 2) produce realistic images without generator training, and 3) give users intuitive control over the generation process.
arXiv Detail & Related papers (2021-04-13T02:00:24Z) - Towards End-to-End Neural Face Authentication in the Wild - Quantifying
and Compensating for Directional Lighting Effects [2.4493299476776778]
This work examines the effects of directional lighting on a State-of-Art(SoA) neural face recognizer.
Top lighting and its variants are found to have minimal effect on accuracy, while bottom-left or bottom-right directional lighting has the most pronounced effects.
arXiv Detail & Related papers (2021-04-08T15:58:09Z) - Identity-Aware CycleGAN for Face Photo-Sketch Synthesis and Recognition [61.87842307164351]
We first propose an Identity-Aware CycleGAN (IACycleGAN) model that applies a new perceptual loss to supervise the image generation network.
It improves CycleGAN on photo-sketch synthesis by paying more attention to the synthesis of key facial regions, such as eyes and nose.
We develop a mutual optimization procedure between the synthesis model and the recognition model, which iteratively synthesizes better images by IACycleGAN.
arXiv Detail & Related papers (2021-03-30T01:30:08Z) - Pixel-based Facial Expression Synthesis [1.7056768055368383]
We propose a pixel-based facial expression synthesis method in which each output pixel observes only one input pixel.
The proposed model is two orders of magnitude smaller which makes it suitable for deployment on resource-constrained devices.
arXiv Detail & Related papers (2020-10-27T16:00:45Z) - Exposing GAN-generated Faces Using Inconsistent Corneal Specular
Highlights [42.83346543247565]
We show that GAN synthesized faces can be exposed with the inconsistent corneal specular highlights between two eyes.
The inconsistency is caused by the lack of physical/physiological constraints in the GAN models.
arXiv Detail & Related papers (2020-09-24T19:43:16Z) - InterFaceGAN: Interpreting the Disentangled Face Representation Learned
by GANs [73.27299786083424]
We propose a framework called InterFaceGAN to interpret the disentangled face representation learned by state-of-the-art GAN models.
We first find that GANs learn various semantics in some linear subspaces of the latent space.
We then conduct a detailed study on the correlation between different semantics and manage to better disentangle them via subspace projection.
arXiv Detail & Related papers (2020-05-18T18:01:22Z) - Joint Deep Learning of Facial Expression Synthesis and Recognition [97.19528464266824]
We propose a novel joint deep learning of facial expression synthesis and recognition method for effective FER.
The proposed method involves a two-stage learning procedure. Firstly, a facial expression synthesis generative adversarial network (FESGAN) is pre-trained to generate facial images with different facial expressions.
In order to alleviate the problem of data bias between the real images and the synthetic images, we propose an intra-class loss with a novel real data-guided back-propagation (RDBP) algorithm.
arXiv Detail & Related papers (2020-02-06T10:56:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.