Multi-task Cross Attention Network in Facial Behavior Analysis
- URL: http://arxiv.org/abs/2207.10293v1
- Date: Thu, 21 Jul 2022 04:07:07 GMT
- Title: Multi-task Cross Attention Network in Facial Behavior Analysis
- Authors: Dang-Khanh Nguyen, Sudarshan Pant, Ngoc-Huynh Ho, Guee-Sang Lee,
Soo-Huyng Kim, Hyung-Jeong Yang
- Abstract summary: We present our solution for the Multi-Task Learning challenge of the Affective Behavior Analysis in-the-wild competition.
The challenge is a combination of three tasks: action unit detection, facial expression recognition and valance-arousal estimation.
We introduce a cross-attentive module to improve multi-task learning performance.
- Score: 7.910908058662372
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Facial behavior analysis is a broad topic with various categories such as
facial emotion recognition, age and gender recognition, ... Many studies focus
on individual tasks while the multi-task learning approach is still open and
requires more research. In this paper, we present our solution and experiment
result for the Multi-Task Learning challenge of the Affective Behavior Analysis
in-the-wild competition. The challenge is a combination of three tasks: action
unit detection, facial expression recognition and valance-arousal estimation.
To address this challenge, we introduce a cross-attentive module to improve
multi-task learning performance. Additionally, a facial graph is applied to
capture the association among action units. As a result, we achieve the
evaluation measure of 1.24 on the validation data provided by the organizers,
which is better than the baseline result of 0.30.
Related papers
- Affective Behaviour Analysis via Progressive Learning [23.455163723584427]
We present our methods and experimental results for the two competition tracks.
We train a Masked-Auto in a self-supervised manner to attain high-quality facial features.
We utilize curriculum learning to transition the model from recognizing single expressions to recognizing compound expressions.
arXiv Detail & Related papers (2024-07-24T02:24:21Z) - Affective Behavior Analysis using Task-adaptive and AU-assisted Graph Network [18.304164382834617]
We present our solution and experiment result for the Multi-Task Learning Challenge of the 7th Affective Behavior Analysis in-the-wild(ABAW7) Competition.
This challenge consists of three tasks: action unit detection, facial expression recognition, and valance-arousal estimation.
arXiv Detail & Related papers (2024-07-16T12:33:22Z) - Multi-Task Learning for Affect Analysis [0.0]
This project investigates two primary approaches: uni-task solutions and a multi-task approach to the same problems.
The project utilizes existing a neural network architecture, adapting it for multi-task learning by modifying output layers and loss functions.
The research aspires to contribute to the burgeoning field of affective computing, with applications spanning healthcare, marketing, and human-computer interaction.
arXiv Detail & Related papers (2024-06-30T12:36:37Z) - Task-adaptive Q-Face [75.15668556061772]
We propose a novel task-adaptive multi-task face analysis method named as Q-Face.
Q-Face simultaneously performs multiple face analysis tasks with a unified model.
Our method achieves state-of-the-art performance on face expression recognition, action unit detection, face attribute analysis, age estimation, and face pose estimation.
arXiv Detail & Related papers (2024-05-15T03:13:11Z) - Human Reaction Intensity Estimation with Ensemble of Multi-task Networks [2.6432771146480283]
"Emotional Reaction Intensity" (ERI) is an important topic in the facial expression recognition task.
We propose a multi-emotional task learning-based approach and present preliminary results for the ERI challenge introduced in the 5th affective behavior analysis in-the-wild (ABAW) competition.
arXiv Detail & Related papers (2023-03-16T11:35:59Z) - Task Formulation Matters When Learning Continually: A Case Study in
Visual Question Answering [58.82325933356066]
Continual learning aims to train a model incrementally on a sequence of tasks without forgetting previous knowledge.
We present a detailed study of how different settings affect performance for Visual Question Answering.
arXiv Detail & Related papers (2022-09-30T19:12:58Z) - Distribution Matching for Heterogeneous Multi-Task Learning: a
Large-scale Face Study [75.42182503265056]
Multi-Task Learning has emerged as a methodology in which multiple tasks are jointly learned by a shared learning algorithm.
We deal with heterogeneous MTL, simultaneously addressing detection, classification & regression problems.
We build FaceBehaviorNet, the first framework for large-scale face analysis, by jointly learning all facial behavior tasks.
arXiv Detail & Related papers (2021-05-08T22:26:52Z) - Pre-training strategies and datasets for facial representation learning [58.8289362536262]
We show how to find a universal face representation that can be adapted to several facial analysis tasks and datasets.
We systematically investigate two ways of large-scale representation learning applied to faces: supervised and unsupervised pre-training.
Our main two findings are: Unsupervised pre-training on completely in-the-wild, uncurated data provides consistent and, in some cases, significant accuracy improvements.
arXiv Detail & Related papers (2021-03-30T17:57:25Z) - Affect Analysis in-the-wild: Valence-Arousal, Expressions, Action Units
and a Unified Framework [83.21732533130846]
The paper focuses on large in-the-wild databases, i.e., Aff-Wild and Aff-Wild2.
It presents the design of two classes of deep neural networks trained with these databases.
A novel multi-task and holistic framework is presented which is able to jointly learn and effectively generalize and perform affect recognition.
arXiv Detail & Related papers (2021-03-29T17:36:20Z) - LEMMA: A Multi-view Dataset for Learning Multi-agent Multi-task
Activities [119.88381048477854]
We introduce the LEMMA dataset to provide a single home to address missing dimensions with meticulously designed settings.
We densely annotate the atomic-actions with human-object interactions to provide ground-truths of the compositionality, scheduling, and assignment of daily activities.
We hope this effort would drive the machine vision community to examine goal-directed human activities and further study the task scheduling and assignment in the real world.
arXiv Detail & Related papers (2020-07-31T00:13:54Z) - Deep Multi-task Learning for Facial Expression Recognition and Synthesis
Based on Selective Feature Sharing [28.178390846446938]
We propose a novel selective feature-sharing method, and establish a multi-task network for facial expression recognition and facial expression synthesis.
The proposed method can effectively transfer beneficial features between different tasks, while filtering out useless and harmful information.
Experimental results show that the proposed method achieves state-of-the-art performance on those commonly used facial expression recognition benchmarks.
arXiv Detail & Related papers (2020-07-09T02:29:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.