Affect Analysis in-the-wild: Valence-Arousal, Expressions, Action Units
and a Unified Framework
- URL: http://arxiv.org/abs/2103.15792v1
- Date: Mon, 29 Mar 2021 17:36:20 GMT
- Title: Affect Analysis in-the-wild: Valence-Arousal, Expressions, Action Units
and a Unified Framework
- Authors: Dimitrios Kollias and Stefanos Zafeiriou
- Abstract summary: The paper focuses on large in-the-wild databases, i.e., Aff-Wild and Aff-Wild2.
It presents the design of two classes of deep neural networks trained with these databases.
A novel multi-task and holistic framework is presented which is able to jointly learn and effectively generalize and perform affect recognition.
- Score: 83.21732533130846
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Affect recognition based on subjects' facial expressions has been a topic of
major research in the attempt to generate machines that can understand the way
subjects feel, act and react. In the past, due to the unavailability of large
amounts of data captured in real-life situations, research has mainly focused
on controlled environments. However, recently, social media and platforms have
been widely used. Moreover, deep learning has emerged as a means to solve
visual analysis and recognition problems. This paper exploits these advances
and presents significant contributions for affect analysis and recognition
in-the-wild. Affect analysis and recognition can be seen as a dual knowledge
generation problem, involving: i) creation of new, large and rich in-the-wild
databases and ii) design and training of novel deep neural architectures that
are able to analyse affect over these databases and to successfully generalise
their performance on other datasets. The paper focuses on large in-the-wild
databases, i.e., Aff-Wild and Aff-Wild2 and presents the design of two classes
of deep neural networks trained with these databases. The first class refers to
uni-task affect recognition, focusing on prediction of the valence and arousal
dimensional variables. The second class refers to estimation of all main
behavior tasks, i.e. valence-arousal prediction; categorical emotion
classification in seven basic facial expressions; facial Action Unit detection.
A novel multi-task and holistic framework is presented which is able to jointly
learn and effectively generalize and perform affect recognition over all
existing in-the-wild databases. Large experimental studies illustrate the
achieved performance improvement over the existing state-of-the-art in affect
recognition.
Related papers
- Balancing the Scales: Enhancing Fairness in Facial Expression Recognition with Latent Alignment [5.784550537553534]
This workleverages representation learning based on latent spaces to mitigate bias in facial expression recognition systems.
It also enhances a deep learning model's fairness and overall accuracy.
arXiv Detail & Related papers (2024-10-25T10:03:10Z) - Object recognition in primates: What can early visual areas contribute? [0.0]
We investigate how signals carried by early visual processing areas could be used for object recognition in the periphery.
Models of V1 simple or complex cells could provide quite reliable information, resulting in performance better than 80% in realistic scenarios.
We propose that object recognition should be seen as a parallel process, with high-accuracy foveal modules operating in parallel with lower-accuracy and faster modules that can operate across the visual field.
arXiv Detail & Related papers (2024-07-05T18:57:09Z) - Investigating Human-Identifiable Features Hidden in Adversarial
Perturbations [54.39726653562144]
Our study explores up to five attack algorithms across three datasets.
We identify human-identifiable features in adversarial perturbations.
Using pixel-level annotations, we extract such features and demonstrate their ability to compromise target models.
arXiv Detail & Related papers (2023-09-28T22:31:29Z) - CIAO! A Contrastive Adaptation Mechanism for Non-Universal Facial
Expression Recognition [80.07590100872548]
We propose Contrastive Inhibitory Adaptati On (CIAO), a mechanism that adapts the last layer of facial encoders to depict specific affective characteristics on different datasets.
CIAO presents an improvement in facial expression recognition performance over six different datasets with very unique affective representations.
arXiv Detail & Related papers (2022-08-10T15:46:05Z) - Causal Reasoning Meets Visual Representation Learning: A Prospective
Study [117.08431221482638]
Lack of interpretability, robustness, and out-of-distribution generalization are becoming the challenges of the existing visual models.
Inspired by the strong inference ability of human-level agents, recent years have witnessed great effort in developing causal reasoning paradigms.
This paper aims to provide a comprehensive overview of this emerging field, attract attention, encourage discussions, bring to the forefront the urgency of developing novel causal reasoning methods.
arXiv Detail & Related papers (2022-04-26T02:22:28Z) - Fine-Grained Image Analysis with Deep Learning: A Survey [146.22351342315233]
Fine-grained image analysis (FGIA) is a longstanding and fundamental problem in computer vision and pattern recognition.
This paper attempts to re-define and broaden the field of FGIA by consolidating two fundamental fine-grained research areas -- fine-grained image recognition and fine-grained image retrieval.
arXiv Detail & Related papers (2021-11-11T09:43:56Z) - Towards Unbiased Visual Emotion Recognition via Causal Intervention [63.74095927462]
We propose a novel Emotion Recognition Network (IERN) to alleviate the negative effects brought by the dataset bias.
A series of designed tests validate the effectiveness of IERN, and experiments on three emotion benchmarks demonstrate that IERN outperforms other state-of-the-art approaches.
arXiv Detail & Related papers (2021-07-26T10:40:59Z) - Prior Aided Streaming Network for Multi-task Affective Recognitionat the
2nd ABAW2 Competition [9.188777864190204]
We introduce our submission to the 2nd Affective Behavior Analysis in-the-wild (ABAW2) Competition.
In dealing with different emotion representations, we propose a multi-task streaming network.
We leverage an advanced facial expression embedding as prior knowledge.
arXiv Detail & Related papers (2021-07-08T09:35:08Z) - Understanding top-down attention using task-oriented ablation design [0.22940141855172028]
Top-down attention allows neural networks, both artificial and biological, to focus on the information most relevant for a given task.
We aim to answer this with a computational experiment based on a general framework called task-oriented ablation design.
We compare the performance of two neural networks, one with top-down attention and one without.
arXiv Detail & Related papers (2021-06-08T21:01:47Z) - A Multi-resolution Approach to Expression Recognition in the Wild [9.118706387430883]
We propose a multi-resolution approach to solve the Facial Expression Recognition task.
We ground our intuition on the observation that often faces images are acquired at different resolutions.
To our aim, we use a ResNet-like architecture, equipped with Squeeze-and-Excitation blocks, trained on the Affect-in-the-Wild 2 dataset.
arXiv Detail & Related papers (2021-03-09T21:21:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.