When Facial Expression Recognition Meets Few-Shot Learning: A Joint and
Alternate Learning Framework
- URL: http://arxiv.org/abs/2201.06781v1
- Date: Tue, 18 Jan 2022 07:24:12 GMT
- Title: When Facial Expression Recognition Meets Few-Shot Learning: A Joint and
Alternate Learning Framework
- Authors: Xinyi Zou, Yan Yan, Jing-Hao Xue, Si Chen, Hanzi Wang
- Abstract summary: We propose an Emotion Guided Similarity Network (EGS-Net) to address the diversity of human emotions in practical scenarios.
EGS-Net consists of an emotion branch and a similarity branch, based on a two-stage learning framework.
Experimental results on both in-the-lab and in-the-wild compound expression datasets demonstrate the superiority of our proposed method against several state-of-the-art methods.
- Score: 60.51225419301642
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Human emotions involve basic and compound facial expressions. However,
current research on facial expression recognition (FER) mainly focuses on basic
expressions, and thus fails to address the diversity of human emotions in
practical scenarios. Meanwhile, existing work on compound FER relies heavily on
abundant labeled compound expression training data, which are often laboriously
collected under the professional instruction of psychology. In this paper, we
study compound FER in the cross-domain few-shot learning setting, where only a
few images of novel classes from the target domain are required as a reference.
In particular, we aim to identify unseen compound expressions with the model
trained on easily accessible basic expression datasets. To alleviate the
problem of limited base classes in our FER task, we propose a novel Emotion
Guided Similarity Network (EGS-Net), consisting of an emotion branch and a
similarity branch, based on a two-stage learning framework. Specifically, in
the first stage, the similarity branch is jointly trained with the emotion
branch in a multi-task fashion. With the regularization of the emotion branch,
we prevent the similarity branch from overfitting to sampled base classes that
are highly overlapped across different episodes. In the second stage, the
emotion branch and the similarity branch play a "two-student game" to
alternately learn from each other, thereby further improving the inference
ability of the similarity branch on unseen compound expressions. Experimental
results on both in-the-lab and in-the-wild compound expression datasets
demonstrate the superiority of our proposed method against several
state-of-the-art methods.
Related papers
- Two in One Go: Single-stage Emotion Recognition with Decoupled Subject-context Transformer [78.35816158511523]
We present a single-stage emotion recognition approach, employing a Decoupled Subject-Context Transformer (DSCT) for simultaneous subject localization and emotion classification.
We evaluate our single-stage framework on two widely used context-aware emotion recognition datasets, CAER-S and EMOTIC.
arXiv Detail & Related papers (2024-04-26T07:30:32Z) - Multi-Cue Adaptive Emotion Recognition Network [4.570705738465714]
We propose a new deep learning approach for emotion recognition based on adaptive multi-cues.
We compare the proposed approach with the state-of-art approaches in the CAER-S dataset.
arXiv Detail & Related papers (2021-11-03T15:08:55Z) - A Circular-Structured Representation for Visual Emotion Distribution
Learning [82.89776298753661]
We propose a well-grounded circular-structured representation to utilize the prior knowledge for visual emotion distribution learning.
To be specific, we first construct an Emotion Circle to unify any emotional state within it.
On the proposed Emotion Circle, each emotion distribution is represented with an emotion vector, which is defined with three attributes.
arXiv Detail & Related papers (2021-06-23T14:53:27Z) - A Generalized Zero-Shot Framework for Emotion Recognition from Body
Gestures [5.331671302839567]
We introduce a Generalized Zero-Shot Learning (GZSL) framework to infer the emotional state of new body gestures.
The framework is significantly superior to the traditional method of emotion classification and state-of-the-art zero-shot learning methods.
arXiv Detail & Related papers (2020-10-13T13:16:38Z) - Modality-Transferable Emotion Embeddings for Low-Resource Multimodal
Emotion Recognition [55.44502358463217]
We propose a modality-transferable model with emotion embeddings to tackle the aforementioned issues.
Our model achieves state-of-the-art performance on most of the emotion categories.
Our model also outperforms existing baselines in the zero-shot and few-shot scenarios for unseen emotions.
arXiv Detail & Related papers (2020-09-21T06:10:39Z) - RAF-AU Database: In-the-Wild Facial Expressions with Subjective Emotion
Judgement and Objective AU Annotations [36.93475723886278]
We develop a RAF-AU database that employs a sign-based (i.e., AUs) and judgement-based (i.e., perceived emotion) approach to annotating blended facial expressions in the wild.
We also conduct a preliminary investigation of which key AUs contribute most to a perceived emotion, and the relationship between AUs and facial expressions.
arXiv Detail & Related papers (2020-08-12T09:29:16Z) - Symbiotic Adversarial Learning for Attribute-based Person Search [86.7506832053208]
We present a symbiotic adversarial learning framework, called SAL.Two GANs sit at the base of the framework in a symbiotic learning scheme.
Specifically, two different types of generative adversarial networks learn collaboratively throughout the training process.
arXiv Detail & Related papers (2020-07-19T07:24:45Z) - Learning to Augment Expressions for Few-shot Fine-grained Facial
Expression Recognition [98.83578105374535]
We present a novel Fine-grained Facial Expression Database - F2ED.
It includes more than 200k images with 54 facial expressions from 119 persons.
Considering the phenomenon of uneven data distribution and lack of samples is common in real-world scenarios, we evaluate several tasks of few-shot expression learning.
We propose a unified task-driven framework - Compositional Generative Adversarial Network (Comp-GAN) learning to synthesize facial images.
arXiv Detail & Related papers (2020-01-17T03:26:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.