Continual Facial Expression Recognition: A Benchmark
- URL: http://arxiv.org/abs/2305.06448v1
- Date: Wed, 10 May 2023 20:35:38 GMT
- Title: Continual Facial Expression Recognition: A Benchmark
- Authors: Nikhil Churamani, Tolga Dimlioglu, German I. Parisi and Hatice Gunes
- Abstract summary: This work presents the Continual Facial Expression Recognition (ConFER) benchmark that evaluates popular CL techniques on FER tasks.
It presents a comparative analysis of several CL-based approaches on popular FER datasets such as CK+, RAF-DB, and AffectNet.
CL techniques, under different learning settings, are shown to achieve state-of-the-art (SOTA) performance across several datasets.
- Score: 3.181579197770883
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding human affective behaviour, especially in the dynamics of
real-world settings, requires Facial Expression Recognition (FER) models to
continuously adapt to individual differences in user expression, contextual
attributions, and the environment. Current (deep) Machine Learning (ML)-based
FER approaches pre-trained in isolation on benchmark datasets fail to capture
the nuances of real-world interactions where data is available only
incrementally, acquired by the agent or robot during interactions. New learning
comes at the cost of previous knowledge, resulting in catastrophic forgetting.
Lifelong or Continual Learning (CL), on the other hand, enables adaptability in
agents by being sensitive to changing data distributions, integrating new
information without interfering with previously learnt knowledge. Positing CL
as an effective learning paradigm for FER, this work presents the Continual
Facial Expression Recognition (ConFER) benchmark that evaluates popular CL
techniques on FER tasks. It presents a comparative analysis of several CL-based
approaches on popular FER datasets such as CK+, RAF-DB, and AffectNet and
present strategies for a successful implementation of ConFER for Affective
Computing (AC) research. CL techniques, under different learning settings, are
shown to achieve state-of-the-art (SOTA) performance across several datasets,
thus motivating a discussion on the benefits of applying CL principles towards
human behaviour understanding, particularly from facial expressions, as well
the challenges entailed.
Related papers
- Investigating the Pre-Training Dynamics of In-Context Learning: Task Recognition vs. Task Learning [99.05401042153214]
In-context learning (ICL) is potentially attributed to two major abilities: task recognition (TR) and task learning (TL)
We take the first step by examining the pre-training dynamics of the emergence of ICL.
We propose a simple yet effective method to better integrate these two abilities for ICL at inference time.
arXiv Detail & Related papers (2024-06-20T06:37:47Z) - Facial Affective Behavior Analysis with Instruction Tuning [58.332959295770614]
Facial affective behavior analysis (FABA) is crucial for understanding human mental states from images.
Traditional approaches primarily deploy models to discriminate among discrete emotion categories, and lack the fine granularity and reasoning capability for complex facial behaviors.
We introduce an instruction-following dataset for two FABA tasks, emotion and action unit recognition, and a benchmark FABA-Bench with a new metric considering both recognition and generation ability.
We also introduce a facial prior expert module with face structure knowledge and a low-rank adaptation module into pre-trained MLLM.
arXiv Detail & Related papers (2024-04-07T19:23:28Z) - Emotic Masked Autoencoder with Attention Fusion for Facial Expression Recognition [1.4374467687356276]
This paper presents an innovative approach integrating the MAE-Face self-supervised learning (SSL) method and multi-view Fusion Attention mechanism for expression classification.
We suggest easy-to-implement and no-training frameworks aimed at highlighting key facial features to determine if such features can serve as guides for the model.
The efficacy of this method is validated by improvements in model performance on the Aff-wild2 dataset.
arXiv Detail & Related papers (2024-03-19T16:21:47Z) - Cluster-level pseudo-labelling for source-free cross-domain facial
expression recognition [94.56304526014875]
We propose the first Source-Free Unsupervised Domain Adaptation (SFUDA) method for Facial Expression Recognition (FER)
Our method exploits self-supervised pretraining to learn good feature representations from the target data.
We validate the effectiveness of our method in four adaptation setups, proving that it consistently outperforms existing SFUDA methods when applied to FER.
arXiv Detail & Related papers (2022-10-11T08:24:50Z) - Learning Deep Representations via Contrastive Learning for Instance
Retrieval [11.736450745549792]
This paper makes the first attempt that tackles the problem using instance-discrimination based contrastive learning (CL)
In this work, we approach this problem by exploring the capability of deriving discriminative representations from pre-trained and fine-tuned CL models.
arXiv Detail & Related papers (2022-09-28T04:36:34Z) - Using Representation Expressiveness and Learnability to Evaluate
Self-Supervised Learning Methods [61.49061000562676]
We introduce Cluster Learnability (CL) to assess learnability.
CL is measured in terms of the performance of a KNN trained to predict labels obtained by clustering the representations with K-means.
We find that CL better correlates with in-distribution model performance than other competing recent evaluation schemes.
arXiv Detail & Related papers (2022-06-02T19:05:13Z) - Confounder Identification-free Causal Visual Feature Learning [84.28462256571822]
We propose a novel Confounder Identification-free Causal Visual Feature Learning (CICF) method, which obviates the need for identifying confounders.
CICF models the interventions among different samples based on front-door criterion, and then approximates the global-scope intervening effect upon the instance-level interventions.
We uncover the relation between CICF and the popular meta-learning strategy MAML, and provide an interpretation of why MAML works from the theoretical perspective.
arXiv Detail & Related papers (2021-11-26T10:57:47Z) - Towards Fair Affective Robotics: Continual Learning for Mitigating Bias
in Facial Expression and Action Unit Recognition [5.478764356647437]
We propose Continual Learning (CL) as an effective strategy to enhance fairness in Facial Expression Recognition (FER) systems.
We compare different state-of-the-art bias mitigation approaches with CL-based strategies for fairness on expression recognition and Action Unit (AU) detection tasks.
Our experiments show that CL-based methods, on average, outperform popular bias mitigation techniques.
arXiv Detail & Related papers (2021-03-15T18:36:14Z) - ORDisCo: Effective and Efficient Usage of Incremental Unlabeled Data for
Semi-supervised Continual Learning [52.831894583501395]
Continual learning assumes the incoming data are fully labeled, which might not be applicable in real applications.
We propose deep Online Replay with Discriminator Consistency (ORDisCo) to interdependently learn a classifier with a conditional generative adversarial network (GAN)
We show ORDisCo achieves significant performance improvement on various semi-supervised learning benchmark datasets for SSCL.
arXiv Detail & Related papers (2021-01-02T09:04:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.