Contrastive Learning of View-Invariant Representations for Facial
Expressions Recognition
- URL: http://arxiv.org/abs/2311.06852v1
- Date: Sun, 12 Nov 2023 14:05:09 GMT
- Title: Contrastive Learning of View-Invariant Representations for Facial
Expressions Recognition
- Authors: Shuvendu Roy, Ali Etemad
- Abstract summary: We propose ViewFX, a novel view-invariant FER framework based on contrastive learning.
We test the proposed framework on two public multi-view facial expression recognition datasets.
- Score: 27.75143621836449
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Although there has been much progress in the area of facial expression
recognition (FER), most existing methods suffer when presented with images that
have been captured from viewing angles that are non-frontal and substantially
different from those used in the training process. In this paper, we propose
ViewFX, a novel view-invariant FER framework based on contrastive learning,
capable of accurately classifying facial expressions regardless of the input
viewing angles during inference. ViewFX learns view-invariant features of
expression using a proposed self-supervised contrastive loss which brings
together different views of the same subject with a particular expression in
the embedding space. We also introduce a supervised contrastive loss to push
the learnt view-invariant features of each expression away from other
expressions. Since facial expressions are often distinguished with very subtle
differences in the learned feature space, we incorporate the Barlow twins loss
to reduce the redundancy and correlations of the representations in the learned
representations. The proposed method is a substantial extension of our
previously proposed CL-MEx, which only had a self-supervised loss. We test the
proposed framework on two public multi-view facial expression recognition
datasets, KDEF and DDCF. The experiments demonstrate that our approach
outperforms previous works in the area and sets a new state-of-the-art for both
datasets while showing considerably less sensitivity to challenging angles and
the number of output labels used for training. We also perform detailed
sensitivity and ablation experiments to evaluate the impact of different
components of our model as well as its sensitivity to different parameters.
Related papers
- Knowledge-Enhanced Facial Expression Recognition with Emotional-to-Neutral Transformation [66.53435569574135]
Existing facial expression recognition methods typically fine-tune a pre-trained visual encoder using discrete labels.
We observe that the rich knowledge in text embeddings, generated by vision-language models, is a promising alternative for learning discriminative facial expression representations.
We propose a novel knowledge-enhanced FER method with an emotional-to-neutral transformation.
arXiv Detail & Related papers (2024-09-13T07:28:57Z) - GaFET: Learning Geometry-aware Facial Expression Translation from
In-The-Wild Images [55.431697263581626]
We introduce a novel Geometry-aware Facial Expression Translation framework, which is based on parametric 3D facial representations and can stably decoupled expression.
We achieve higher-quality and more accurate facial expression transfer results compared to state-of-the-art methods, and demonstrate applicability of various poses and complex textures.
arXiv Detail & Related papers (2023-08-07T09:03:35Z) - Disentangling Identity and Pose for Facial Expression Recognition [54.50747989860957]
We propose an identity and pose disentangled facial expression recognition (IPD-FER) model to learn more discriminative feature representation.
For identity encoder, a well pre-trained face recognition model is utilized and fixed during training, which alleviates the restriction on specific expression training data.
By comparing the difference between synthesized neutral and expressional images of the same individual, the expression component is further disentangled from identity and pose.
arXiv Detail & Related papers (2022-08-17T06:48:13Z) - CIAO! A Contrastive Adaptation Mechanism for Non-Universal Facial
Expression Recognition [80.07590100872548]
We propose Contrastive Inhibitory Adaptati On (CIAO), a mechanism that adapts the last layer of facial encoders to depict specific affective characteristics on different datasets.
CIAO presents an improvement in facial expression recognition performance over six different datasets with very unique affective representations.
arXiv Detail & Related papers (2022-08-10T15:46:05Z) - Visual Perturbation-aware Collaborative Learning for Overcoming the
Language Prior Problem [60.0878532426877]
We propose a novel collaborative learning scheme from the viewpoint of visual perturbation calibration.
Specifically, we devise a visual controller to construct two sorts of curated images with different perturbation extents.
The experimental results on two diagnostic VQA-CP benchmark datasets evidently demonstrate its effectiveness.
arXiv Detail & Related papers (2022-07-24T23:50:52Z) - Dual Contrastive Learning for General Face Forgery Detection [64.41970626226221]
We propose a novel face forgery detection framework, named Dual Contrastive Learning (DCL), which constructs positive and negative paired data.
To explore the essential discrepancies, Intra-Instance Contrastive Learning (Intra-ICL) is introduced to focus on the local content inconsistencies prevalent in the forged faces.
arXiv Detail & Related papers (2021-12-27T05:44:40Z) - Self-supervised Contrastive Learning of Multi-view Facial Expressions [9.949781365631557]
Facial expression recognition (FER) has emerged as an important component of human-computer interaction systems.
We propose Contrastive Learning of Multi-view facial Expressions (CL-MEx) to exploit facial images captured simultaneously from different angles towards FER.
arXiv Detail & Related papers (2021-08-15T11:23:34Z) - Exploiting Emotional Dependencies with Graph Convolutional Networks for
Facial Expression Recognition [31.40575057347465]
This paper proposes a novel multi-task learning framework to recognize facial expressions in-the-wild.
A shared feature representation is learned for both discrete and continuous recognition in a MTL setting.
The results of our experiments show that our method outperforms the current state-of-the-art methods on discrete FER.
arXiv Detail & Related papers (2021-06-07T10:20:05Z) - Learning Disentangled Expression Representations from Facial Images [2.2509387878255818]
We use a formulation of the adversarial loss to learn disentangled representations for face images.
The used model facilitates learning on single-task datasets and improves the state-of-the-art in expression recognition with an accuracy of60.53%.
arXiv Detail & Related papers (2020-08-16T21:23:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.