Generative Adversarial Stacked Autoencoders for Facial Pose
Normalization and Emotion Recognition
- URL: http://arxiv.org/abs/2007.09790v1
- Date: Sun, 19 Jul 2020 21:47:16 GMT
- Title: Generative Adversarial Stacked Autoencoders for Facial Pose
Normalization and Emotion Recognition
- Authors: Ariel Ruiz-Garcia, Vasile Palade, Mark Elshaw, Mariette Awad
- Abstract summary: We propose a Generative Adversarial Stacked Autoencoder that learns to map facial expressions.
We report state-of-the-art performance on several facial emotion recognition corpora, including one collected in the wild.
- Score: 4.620526905329234
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In this work, we propose a novel Generative Adversarial Stacked Autoencoder
that learns to map facial expressions, with up to plus or minus 60 degrees, to
an illumination invariant facial representation of 0 degrees. We accomplish
this by using a novel convolutional layer that exploits both local and global
spatial information, and a convolutional layer with a reduced number of
parameters that exploits facial symmetry. Furthermore, we introduce a
generative adversarial gradual greedy layer-wise learning algorithm designed to
train Adversarial Autoencoders in an efficient and incremental manner. We
demonstrate the efficiency of our method and report state-of-the-art
performance on several facial emotion recognition corpora, including one
collected in the wild.
Related papers
- ID-Guard: A Universal Framework for Combating Facial Manipulation via Breaking Identification [60.73617868629575]
misuse of deep learning-based facial manipulation poses a potential threat to civil rights.
To prevent this fraud at its source, proactive defense technology was proposed to disrupt the manipulation process.
We propose a novel universal framework for combating facial manipulation, called ID-Guard.
arXiv Detail & Related papers (2024-09-20T09:30:08Z) - G2Face: High-Fidelity Reversible Face Anonymization via Generative and Geometric Priors [71.69161292330504]
Reversible face anonymization seeks to replace sensitive identity information in facial images with synthesized alternatives.
This paper introduces Gtextsuperscript2Face, which leverages both generative and geometric priors to enhance identity manipulation.
Our method outperforms existing state-of-the-art techniques in face anonymization and recovery, while preserving high data utility.
arXiv Detail & Related papers (2024-08-18T12:36:47Z) - Masked Face Recognition with Generative-to-Discriminative Representations [29.035270415311427]
We propose a unified deep network to learn generative-to-discriminative representations for facilitating masked face recognition.
First, we leverage a generative encoder pretrained for face inpainting and finetune it to represent masked faces into category-aware descriptors.
We incorporate a multi-layer convolutional network as a discriminative reformer and learn it to convert the category-aware descriptors into identity-aware vectors.
arXiv Detail & Related papers (2024-05-27T02:20:55Z) - LCM-Lookahead for Encoder-based Text-to-Image Personalization [82.56471486184252]
We explore the potential of using shortcut-mechanisms to guide the personalization of text-to-image models.
We focus on encoder-based personalization approaches, and demonstrate that by tuning them with a lookahead identity loss, we can achieve higher identity fidelity.
arXiv Detail & Related papers (2024-04-04T17:43:06Z) - E2F-Net: Eyes-to-Face Inpainting via StyleGAN Latent Space [4.110419543591102]
We propose a Generative Adversarial Network (GAN)-based model called Eyes-to-Face Network (E2F-Net)
The proposed approach extracts identity and non-identity features from the periocular region using two dedicated encoders.
We show that our method successfully reconstructs the whole face with high quality, surpassing current techniques.
arXiv Detail & Related papers (2024-03-18T19:11:34Z) - DreamIdentity: Improved Editability for Efficient Face-identity
Preserved Image Generation [69.16517915592063]
We propose a novel face-identity encoder to learn an accurate representation of human faces.
We also propose self-augmented editability learning to enhance the editability of models.
Our methods can generate identity-preserved images under different scenes at a much faster speed.
arXiv Detail & Related papers (2023-07-01T11:01:17Z) - Multi-Domain Norm-referenced Encoding Enables Data Efficient Transfer
Learning of Facial Expression Recognition [62.997667081978825]
We propose a biologically-inspired mechanism for transfer learning in facial expression recognition.
Our proposed architecture provides an explanation for how the human brain might innately recognize facial expressions on varying head shapes.
Our model achieves a classification accuracy of 92.15% on the FERG dataset with extreme data efficiency.
arXiv Detail & Related papers (2023-04-05T09:06:30Z) - Attribute-preserving Face Dataset Anonymization via Latent Code
Optimization [64.4569739006591]
We present a task-agnostic anonymization procedure that directly optimize the images' latent representation in the latent space of a pre-trained GAN.
We demonstrate through a series of experiments that our method is capable of anonymizing the identity of the images whilst -- crucially -- better-preserving the facial attributes.
arXiv Detail & Related papers (2023-03-20T17:34:05Z) - Emotion Transfer Using Vector-Valued Infinite Task Learning [2.588412672658578]
We present a novel style transfer framework building upon infinite task learning and vector-valued reproducing kernel Hilbert spaces.
We instantiate the idea in emotion transfer where the goal is to transform facial images to different target emotions.
arXiv Detail & Related papers (2021-02-09T19:05:56Z) - Head2Head++: Deep Facial Attributes Re-Targeting [6.230979482947681]
We leverage the 3D geometry of faces and Generative Adversarial Networks (GANs) to design a novel deep learning architecture for the task of facial and head reenactment.
We manage to capture the complex non-rigid facial motion from the driving monocular performances and synthesise temporally consistent videos.
Our system performs end-to-end reenactment in nearly real-time speed (18 fps)
arXiv Detail & Related papers (2020-06-17T23:38:37Z) - FaceShifter: Towards High Fidelity And Occlusion Aware Face Swapping [43.236261887752065]
We propose a novel two-stage framework, called FaceShifter, for high fidelity and occlusion aware face swapping.
In its first stage, our framework generates the swapped face in high-fidelity by exploiting and integrating the target attributes thoroughly and adaptively.
To address the challenging facial synthesiss, we append a second stage consisting of a novel Heuristic Error Acknowledging Refinement Network (HEAR-Net)
arXiv Detail & Related papers (2019-12-31T17:57:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.