Pose-disentangled Contrastive Learning for Self-supervised Facial
Representation
- URL: http://arxiv.org/abs/2211.13490v2
- Date: Mon, 8 May 2023 06:37:40 GMT
- Title: Pose-disentangled Contrastive Learning for Self-supervised Facial
Representation
- Authors: Yuanyuan Liu, Wenbin Wang, Yibing Zhan, Shaoze Feng, Kejun Liu, Zhe
Chen
- Abstract summary: We propose a novel Pose-disentangled Contrastive Learning (PCL) method for general self-supervised facial representation.
Our PCL first devises a pose-disentangled decoder (PDD), which disentangles the pose-related features from the face-aware features.
We then introduce a pose-related contrastive learning scheme that learns pose-related information based on data augmentation of the same image.
- Score: 12.677909048435408
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Self-supervised facial representation has recently attracted increasing
attention due to its ability to perform face understanding without relying on
large-scale annotated datasets heavily. However, analytically, current
contrastive-based self-supervised learning (SSL) still performs
unsatisfactorily for learning facial representation. More specifically,
existing contrastive learning (CL) tends to learn pose-invariant features that
cannot depict the pose details of faces, compromising the learning performance.
To conquer the above limitation of CL, we propose a novel Pose-disentangled
Contrastive Learning (PCL) method for general self-supervised facial
representation. Our PCL first devises a pose-disentangled decoder (PDD) with a
delicately designed orthogonalizing regulation, which disentangles the
pose-related features from the face-aware features; therefore, pose-related and
other pose-unrelated facial information could be performed in individual
subnetworks and do not affect each other's training. Furthermore, we introduce
a pose-related contrastive learning scheme that learns pose-related information
based on data augmentation of the same image, which would deliver more
effective face-aware representation for various downstream tasks. We conducted
linear evaluation on four challenging downstream facial understanding tasks,
ie, facial expression recognition, face recognition, AU detection and head pose
estimation. Experimental results demonstrate that our method significantly
outperforms state-of-the-art SSL methods. Code is available at
https://github.com/DreamMr/PCL}{https://github.com/DreamMr/PCL
Related papers
- Self-Supervised Facial Representation Learning with Facial Region
Awareness [13.06996608324306]
Self-supervised pre-training has been proven to be effective in learning transferable representations that benefit various visual tasks.
Recent efforts toward this goal are limited to treating each face image as a whole.
We propose a novel self-supervised facial representation learning framework to learn consistent global and local facial representations.
arXiv Detail & Related papers (2024-03-04T15:48:56Z) - ProS: Facial Omni-Representation Learning via Prototype-based
Self-Distillation [22.30414271893046]
Prototype-based Self-Distillation (ProS) is a novel approach for unsupervised face representation learning.
ProS consists of two vision-transformers (teacher and student models) that are trained with different augmented images.
ProS achieves state-of-the-art performance on various tasks, both in full and few-shot settings.
arXiv Detail & Related papers (2023-11-03T14:10:06Z) - Understanding Pose and Appearance Disentanglement in 3D Human Pose
Estimation [72.50214227616728]
Several methods have proposed to learn image representations in a self-supervised fashion so as to disentangle the appearance information from the pose one.
We study disentanglement from the perspective of the self-supervised network, via diverse image synthesis experiments.
We design an adversarial strategy focusing on generating natural appearance changes of the subject, and against which we could expect a disentangled network to be robust.
arXiv Detail & Related papers (2023-09-20T22:22:21Z) - GaFET: Learning Geometry-aware Facial Expression Translation from
In-The-Wild Images [55.431697263581626]
We introduce a novel Geometry-aware Facial Expression Translation framework, which is based on parametric 3D facial representations and can stably decoupled expression.
We achieve higher-quality and more accurate facial expression transfer results compared to state-of-the-art methods, and demonstrate applicability of various poses and complex textures.
arXiv Detail & Related papers (2023-08-07T09:03:35Z) - CIAO! A Contrastive Adaptation Mechanism for Non-Universal Facial
Expression Recognition [80.07590100872548]
We propose Contrastive Inhibitory Adaptati On (CIAO), a mechanism that adapts the last layer of facial encoders to depict specific affective characteristics on different datasets.
CIAO presents an improvement in facial expression recognition performance over six different datasets with very unique affective representations.
arXiv Detail & Related papers (2022-08-10T15:46:05Z) - General Facial Representation Learning in a Visual-Linguistic Manner [45.92447707178299]
We introduce a framework, called FaRL, for general Facial Representation Learning in a visual-linguistic manner.
We show that FaRL achieves better transfer performance compared with previous pre-trained models.
Our model surpasses the state-of-the-art methods on face analysis tasks including face parsing and face alignment.
arXiv Detail & Related papers (2021-12-06T15:22:05Z) - Attention-guided Progressive Mapping for Profile Face Recognition [12.792576041526289]
Cross pose face recognition remains a significant challenge.
Learning pose-robust features by traversing to the feature space of frontal faces provides an effective and cheap way to alleviate this problem.
arXiv Detail & Related papers (2021-06-27T02:21:41Z) - Learning to Aggregate and Personalize 3D Face from In-the-Wild Photo
Collection [65.92058628082322]
Non-parametric face modeling aims to reconstruct 3D face only from images without shape assumptions.
This paper presents a novel Learning to Aggregate and Personalize framework for unsupervised robust 3D face modeling.
arXiv Detail & Related papers (2021-06-15T03:10:17Z) - Pre-training strategies and datasets for facial representation learning [58.8289362536262]
We show how to find a universal face representation that can be adapted to several facial analysis tasks and datasets.
We systematically investigate two ways of large-scale representation learning applied to faces: supervised and unsupervised pre-training.
Our main two findings are: Unsupervised pre-training on completely in-the-wild, uncurated data provides consistent and, in some cases, significant accuracy improvements.
arXiv Detail & Related papers (2021-03-30T17:57:25Z) - InterFaceGAN: Interpreting the Disentangled Face Representation Learned
by GANs [73.27299786083424]
We propose a framework called InterFaceGAN to interpret the disentangled face representation learned by state-of-the-art GAN models.
We first find that GANs learn various semantics in some linear subspaces of the latent space.
We then conduct a detailed study on the correlation between different semantics and manage to better disentangle them via subspace projection.
arXiv Detail & Related papers (2020-05-18T18:01:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.