LAFS: Landmark-based Facial Self-supervised Learning for Face
Recognition
- URL: http://arxiv.org/abs/2403.08161v1
- Date: Wed, 13 Mar 2024 01:07:55 GMT
- Title: LAFS: Landmark-based Facial Self-supervised Learning for Face
Recognition
- Authors: Zhonglin Sun, Chen Feng, Ioannis Patras, Georgios Tzimiropoulos
- Abstract summary: We focus on learning facial representations that can be adapted to train effective face recognition models.
We explore the learning strategy of unlabeled facial images through self-supervised pretraining.
Our method achieves significant improvement over the state-of-the-art on multiple face recognition benchmarks.
- Score: 37.4550614524874
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work we focus on learning facial representations that can be adapted
to train effective face recognition models, particularly in the absence of
labels. Firstly, compared with existing labelled face datasets, a vastly larger
magnitude of unlabeled faces exists in the real world. We explore the learning
strategy of these unlabeled facial images through self-supervised pretraining
to transfer generalized face recognition performance. Moreover, motivated by
one recent finding, that is, the face saliency area is critical for face
recognition, in contrast to utilizing random cropped blocks of images for
constructing augmentations in pretraining, we utilize patches localized by
extracted facial landmarks. This enables our method - namely LAndmark-based
Facial Self-supervised learning LAFS), to learn key representation that is more
critical for face recognition. We also incorporate two landmark-specific
augmentations which introduce more diversity of landmark information to further
regularize the learning. With learned landmark-based facial representations, we
further adapt the representation for face recognition with regularization
mitigating variations in landmark positions. Our method achieves significant
improvement over the state-of-the-art on multiple face recognition benchmarks,
especially on more challenging few-shot scenarios.
Related papers
- Face Anonymization Made Simple [44.24233169815565]
Current face anonymization techniques often depend on identity loss calculated by face recognition models, which can be inaccurate and unreliable.
In contrast, our approach uses diffusion models with only a reconstruction loss, eliminating the need for facial landmarks or masks.
Our model achieves state-of-the-art performance in three key areas: identity anonymization, facial preservation, and image quality.
arXiv Detail & Related papers (2024-11-01T17:45:21Z) - Self-Supervised Facial Representation Learning with Facial Region
Awareness [13.06996608324306]
Self-supervised pre-training has been proven to be effective in learning transferable representations that benefit various visual tasks.
Recent efforts toward this goal are limited to treating each face image as a whole.
We propose a novel self-supervised facial representation learning framework to learn consistent global and local facial representations.
arXiv Detail & Related papers (2024-03-04T15:48:56Z) - SimFLE: Simple Facial Landmark Encoding for Self-Supervised Facial
Expression Recognition in the Wild [3.4798852684389963]
We propose a self-supervised simple facial landmark encoding (SimFLE) method that can learn effective encoding of facial landmarks.
We introduce novel FaceMAE module for this purpose.
Experimental results on several FER-W benchmarks prove that the proposed SimFLE is superior in facial landmark localization.
arXiv Detail & Related papers (2023-03-14T06:30:55Z) - A Comparative Analysis of the Face Recognition Methods in Video
Surveillance Scenarios [0.0]
This study presents comparative benchmark tables for the state-of-art face recognition methods.
We constructed a video surveillance dataset of face IDs with high age variance, intra-class variance (face make-up, beard, etc.) with native surveillance facial imagery data for evaluation.
On the other hand, this work discovers the best recognition methods for different conditions like non-masked faces, masked faces, and faces with glasses.
arXiv Detail & Related papers (2022-11-05T17:59:18Z) - MorDeephy: Face Morphing Detection Via Fused Classification [0.0]
We introduce a novel deep learning strategy for a single image face morphing detection.
It is directed onto learning the deep facial features, which carry information about the authenticity of these features.
Our method, which we call MorDeephy, achieved the state of the art performance and demonstrated a prominent ability for generalising the task of morphing detection to unseen scenarios.
arXiv Detail & Related papers (2022-08-05T11:39:22Z) - Emotion Separation and Recognition from a Facial Expression by Generating the Poker Face with Vision Transformers [57.1091606948826]
We propose a novel FER model, named Poker Face Vision Transformer or PF-ViT, to address these challenges.
PF-ViT aims to separate and recognize the disturbance-agnostic emotion from a static facial image via generating its corresponding poker face.
PF-ViT utilizes vanilla Vision Transformers, and its components are pre-trained as Masked Autoencoders on a large facial expression dataset.
arXiv Detail & Related papers (2022-07-22T13:39:06Z) - Facial Expressions as a Vulnerability in Face Recognition [73.85525896663371]
This work explores facial expression bias as a security vulnerability of face recognition systems.
We present a comprehensive analysis of how facial expression bias impacts the performance of face recognition technologies.
arXiv Detail & Related papers (2020-11-17T18:12:41Z) - DotFAN: A Domain-transferred Face Augmentation Network for Pose and
Illumination Invariant Face Recognition [94.96686189033869]
We propose a 3D model-assisted domain-transferred face augmentation network (DotFAN)
DotFAN can generate a series of variants of an input face based on the knowledge distilled from existing rich face datasets collected from other domains.
Experiments show that DotFAN is beneficial for augmenting small face datasets to improve their within-class diversity.
arXiv Detail & Related papers (2020-02-23T08:16:34Z) - Dual-Attention GAN for Large-Pose Face Frontalization [59.689836951934694]
We present a novel Dual-Attention Generative Adversarial Network (DA-GAN) for photo-realistic face frontalization.
Specifically, a self-attention-based generator is introduced to integrate local features with their long-range dependencies.
A novel face-attention-based discriminator is applied to emphasize local features of face regions.
arXiv Detail & Related papers (2020-02-17T20:00:56Z) - Investigating the Impact of Inclusion in Face Recognition Training Data
on Individual Face Identification [93.5538147928669]
We audit ArcFace, a state-of-the-art, open source face recognition system, in a large-scale face identification experiment with more than one million distractor images.
We find a Rank-1 face identification accuracy of 79.71% for individuals present in the model's training data and an accuracy of 75.73% for those not present.
arXiv Detail & Related papers (2020-01-09T15:50:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.