Boosting Deep Face Recognition via Disentangling Appearance and Geometry
- URL: http://arxiv.org/abs/2001.04559v1
- Date: Mon, 13 Jan 2020 23:19:58 GMT
- Title: Boosting Deep Face Recognition via Disentangling Appearance and Geometry
- Authors: Ali Dabouei, Fariborz Taherkhani, Sobhan Soleymani, Jeremy Dawson,
Nasser M. Nasrabadi
- Abstract summary: We propose a framework for disentangling the appearance and geometry representations in the face recognition task.
We generate geometrically identical faces by incorporating spatial transformations.
We show that the proposed approach enhances the performance of deep face recognition models.
- Score: 33.196270681809395
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a framework for disentangling the appearance and
geometry representations in the face recognition task. To provide supervision
for this aim, we generate geometrically identical faces by incorporating
spatial transformations. We demonstrate that the proposed approach enhances the
performance of deep face recognition models by assisting the training process
in two ways. First, it enforces the early and intermediate convolutional layers
to learn more representative features that satisfy the properties of
disentangled embeddings. Second, it augments the training set by altering faces
geometrically. Through extensive experiments, we demonstrate that integrating
the proposed approach into state-of-the-art face recognition methods
effectively improves their performance on challenging datasets, such as LFW,
YTF, and MegaFace. Both theoretical and practical aspects of the method are
analyzed rigorously by concerning ablation studies and knowledge transfer
tasks. Furthermore, we show that the knowledge leaned by the proposed method
can favor other face-related tasks, such as attribute prediction.
Related papers
- Emotic Masked Autoencoder with Attention Fusion for Facial Expression Recognition [1.4374467687356276]
This paper presents an innovative approach integrating the MAE-Face self-supervised learning (SSL) method and multi-view Fusion Attention mechanism for expression classification.
We suggest easy-to-implement and no-training frameworks aimed at highlighting key facial features to determine if such features can serve as guides for the model.
The efficacy of this method is validated by improvements in model performance on the Aff-wild2 dataset.
arXiv Detail & Related papers (2024-03-19T16:21:47Z) - Faceptor: A Generalist Model for Face Perception [52.8066001012464]
Faceptor is proposed to adopt a well-designed single-encoder dual-decoder architecture.
Layer-Attention into Faceptor enables the model to adaptively select features from optimal layers to perform the desired tasks.
Our training framework can also be applied to auxiliary supervised learning, significantly improving performance in data-sparse tasks such as age estimation and expression recognition.
arXiv Detail & Related papers (2024-03-14T15:42:31Z) - Fiducial Focus Augmentation for Facial Landmark Detection [4.433764381081446]
We propose a novel image augmentation technique to enhance the model's understanding of facial structures.
We employ a Siamese architecture-based training mechanism with a Deep Canonical Correlation Analysis (DCCA)-based loss.
Our approach outperforms multiple state-of-the-art approaches across various benchmark datasets.
arXiv Detail & Related papers (2024-02-23T01:34:00Z) - Appearance Debiased Gaze Estimation via Stochastic Subject-Wise
Adversarial Learning [33.55397868171977]
Appearance-based gaze estimation has been attracting attention in computer vision, and remarkable improvements have been achieved using various deep learning techniques.
We propose a novel framework: subject-wise gaZE learning (SAZE), which trains a network to generalize the appearance of subjects.
Our experimental results verify the robustness of the method in that it yields state-of-the-art performance, achieving 3.89 and 4.42 on the MPIIGaze and EyeDiap datasets, respectively.
arXiv Detail & Related papers (2024-01-25T00:23:21Z) - Masked Modeling for Self-supervised Representation Learning on Vision
and Beyond [69.64364187449773]
Masked modeling has emerged as a distinctive approach that involves predicting parts of the original data that are proportionally masked during training.
We elaborate on the details of techniques within masked modeling, including diverse masking strategies, recovering targets, network architectures, and more.
We conclude by discussing the limitations of current techniques and point out several potential avenues for advancing masked modeling research.
arXiv Detail & Related papers (2023-12-31T12:03:21Z) - DeepFidelity: Perceptual Forgery Fidelity Assessment for Deepfake
Detection [67.3143177137102]
Deepfake detection refers to detecting artificially generated or edited faces in images or videos.
We propose a novel Deepfake detection framework named DeepFidelity to adaptively distinguish real and fake faces.
arXiv Detail & Related papers (2023-12-07T07:19:45Z) - Enhancing Face Recognition with Latent Space Data Augmentation and
Facial Posture Reconstruction [8.493314424950599]
We propose an approach named the Face Representation Augmentation (FRA) for augmenting face datasets.
FRA is the first method that shifts its focus towards manipulating the face embeddings generated by any face representation learning algorithm.
The proposed method improves the identity classification accuracies by 9.52 %, 10.04 %, and 16.60 %, in comparison with the base models of MagFace, ArcFace, and CosFace, respectively.
arXiv Detail & Related papers (2023-01-27T20:54:58Z) - Recent Progress in Appearance-based Action Recognition [73.6405863243707]
Action recognition is a task to identify various human actions in a video.
Recent appearance-based methods have achieved promising progress towards accurate action recognition.
arXiv Detail & Related papers (2020-11-25T10:18:12Z) - Learning Oracle Attention for High-fidelity Face Completion [121.72704525675047]
We design a comprehensive framework for face completion based on the U-Net structure.
We propose a dual spatial attention module to efficiently learn the correlations between facial textures at multiple scales.
We take the location of the facial components as prior knowledge and impose a multi-discriminator on these regions.
arXiv Detail & Related papers (2020-03-31T01:37:10Z) - Dual-Attention GAN for Large-Pose Face Frontalization [59.689836951934694]
We present a novel Dual-Attention Generative Adversarial Network (DA-GAN) for photo-realistic face frontalization.
Specifically, a self-attention-based generator is introduced to integrate local features with their long-range dependencies.
A novel face-attention-based discriminator is applied to emphasize local features of face regions.
arXiv Detail & Related papers (2020-02-17T20:00:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.