MAFER: a Multi-resolution Approach to Facial Expression Recognition
- URL: http://arxiv.org/abs/2105.02481v1
- Date: Thu, 6 May 2021 07:26:58 GMT
- Title: MAFER: a Multi-resolution Approach to Facial Expression Recognition
- Authors: Fabio Valerio Massoli, Donato Cafarelli, Claudio Gennaro, Giuseppe
Amato, Fabrizio Falchi
- Abstract summary: We propose a two-step learning procedure, named MAFER, to train Deep Learning models tasked with recognizing facial expressions.
A relevant feature of MAFER is that it is task-agnostic, i.e., it can be used complementarily to other objective-related techniques.
- Score: 9.878384185493623
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Emotions play a central role in the social life of every human being, and
their study, which represents a multidisciplinary subject, embraces a great
variety of research fields. Especially concerning the latter, the analysis of
facial expressions represents a very active research area due to its relevance
to human-computer interaction applications. In such a context, Facial
Expression Recognition (FER) is the task of recognizing expressions on human
faces. Typically, face images are acquired by cameras that have, by nature,
different characteristics, such as the output resolution. It has been already
shown in the literature that Deep Learning models applied to face recognition
experience a degradation in their performance when tested against
multi-resolution scenarios. Since the FER task involves analyzing face images
that can be acquired with heterogeneous sources, thus involving images with
different quality, it is plausible to expect that resolution plays an important
role in such a case too. Stemming from such a hypothesis, we prove the benefits
of multi-resolution training for models tasked with recognizing facial
expressions. Hence, we propose a two-step learning procedure, named MAFER, to
train DCNNs to empower them to generate robust predictions across a wide range
of resolutions. A relevant feature of MAFER is that it is task-agnostic, i.e.,
it can be used complementarily to other objective-related techniques. To assess
the effectiveness of the proposed approach, we performed an extensive
experimental campaign on publicly available datasets: \fer{}, \raf{}, and
\oulu{}. For a multi-resolution context, we observe that with our approach,
learning models improve upon the current SotA while reporting comparable
results in fix-resolution contexts. Finally, we analyze the performance of our
models and observe the higher discrimination power of deep features generated
from them.
Related papers
- FFAA: Multimodal Large Language Model based Explainable Open-World Face Forgery Analysis Assistant [59.2438504610849]
We introduce FFAA: Face Forgery Analysis Assistant, consisting of a fine-tuned Multimodal Large Language Model (MLLM) and Multi-answer Intelligent Decision System (MIDS)
Our method not only provides user-friendly and explainable results but also significantly boosts accuracy and robustness compared to previous methods.
arXiv Detail & Related papers (2024-08-19T15:15:20Z) - Task-adaptive Q-Face [75.15668556061772]
We propose a novel task-adaptive multi-task face analysis method named as Q-Face.
Q-Face simultaneously performs multiple face analysis tasks with a unified model.
Our method achieves state-of-the-art performance on face expression recognition, action unit detection, face attribute analysis, age estimation, and face pose estimation.
arXiv Detail & Related papers (2024-05-15T03:13:11Z) - Dynamic Resolution Guidance for Facial Expression Recognition [2.0456513832600884]
This paper introduces a practical method called Dynamic Resolution Guidance for Facial Expression Recognition (DRGFER)
Our framework comprises two main components: the Resolution Recognition Network (RRN) and the Multi-Resolution Adaptation Facial Expression Recognition Network (MRAFER)
The proposed framework exhibits robustness against resolution variations and facial expressions, offering a promising solution for real-world applications.
arXiv Detail & Related papers (2024-04-09T15:02:01Z) - FaceXFormer: A Unified Transformer for Facial Analysis [59.94066615853198]
FaceXformer is an end-to-end unified transformer model for a range of facial analysis tasks.
Our model effectively handles images "in-the-wild," demonstrating its robustness and generalizability across eight different tasks.
arXiv Detail & Related papers (2024-03-19T17:58:04Z) - Explainable Multi-View Deep Networks Methodology for Experimental Physics [0.19574002186090492]
Physical experiments often involve multiple imaging representations, such as X-ray scans and microscopic images.
Deep learning models have been widely used for supervised analysis in these experiments.
Multi-view data has emerged, where each sample is described by views from different angles, sources, or modalities.
There is a lack of proper explainability in multi-view models, which are challenging to explain due to their architectures.
arXiv Detail & Related papers (2023-08-16T08:13:38Z) - CIAO! A Contrastive Adaptation Mechanism for Non-Universal Facial
Expression Recognition [80.07590100872548]
We propose Contrastive Inhibitory Adaptati On (CIAO), a mechanism that adapts the last layer of facial encoders to depict specific affective characteristics on different datasets.
CIAO presents an improvement in facial expression recognition performance over six different datasets with very unique affective representations.
arXiv Detail & Related papers (2022-08-10T15:46:05Z) - LR-to-HR Face Hallucination with an Adversarial Progressive
Attribute-Induced Network [67.64536397027229]
Face super-resolution is a challenging and highly ill-posed problem.
We propose an end-to-end progressive learning framework incorporating facial attributes.
We show that the proposed approach can yield satisfactory face hallucination images outperforming other state-of-the-art approaches.
arXiv Detail & Related papers (2021-09-29T19:50:45Z) - Exploiting Emotional Dependencies with Graph Convolutional Networks for
Facial Expression Recognition [31.40575057347465]
This paper proposes a novel multi-task learning framework to recognize facial expressions in-the-wild.
A shared feature representation is learned for both discrete and continuous recognition in a MTL setting.
The results of our experiments show that our method outperforms the current state-of-the-art methods on discrete FER.
arXiv Detail & Related papers (2021-06-07T10:20:05Z) - A Multi-resolution Approach to Expression Recognition in the Wild [9.118706387430883]
We propose a multi-resolution approach to solve the Facial Expression Recognition task.
We ground our intuition on the observation that often faces images are acquired at different resolutions.
To our aim, we use a ResNet-like architecture, equipped with Squeeze-and-Excitation blocks, trained on the Affect-in-the-Wild 2 dataset.
arXiv Detail & Related papers (2021-03-09T21:21:02Z) - Learning to Augment Expressions for Few-shot Fine-grained Facial
Expression Recognition [98.83578105374535]
We present a novel Fine-grained Facial Expression Database - F2ED.
It includes more than 200k images with 54 facial expressions from 119 persons.
Considering the phenomenon of uneven data distribution and lack of samples is common in real-world scenarios, we evaluate several tasks of few-shot expression learning.
We propose a unified task-driven framework - Compositional Generative Adversarial Network (Comp-GAN) learning to synthesize facial images.
arXiv Detail & Related papers (2020-01-17T03:26:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.