Hybrid CNN-Transformer Model For Facial Affect Recognition In the ABAW4
Challenge
- URL: http://arxiv.org/abs/2207.10201v1
- Date: Wed, 20 Jul 2022 21:38:47 GMT
- Title: Hybrid CNN-Transformer Model For Facial Affect Recognition In the ABAW4
Challenge
- Authors: Lingfeng Wang, Haocheng Li, Chunyin Liu
- Abstract summary: We propose a hybrid CNN-Transformer model for the Multi-Task-Learning (MTL) and Learning from Synthetic Data (LSD) task.
Experimental results on validation dataset shows that our method achieves better performance than baseline model.
- Score: 6.786147929596443
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper describes our submission to the fourth Affective Behavior Analysis
(ABAW) competition. We proposed a hybrid CNN-Transformer model for the
Multi-Task-Learning (MTL) and Learning from Synthetic Data (LSD) task.
Experimental results on validation dataset shows that our method achieves
better performance than baseline model, which verifies that the effectiveness
of proposed network.
Related papers
- Unveiling the Flaws: Exploring Imperfections in Synthetic Data and Mitigation Strategies for Large Language Models [89.88010750772413]
Synthetic data has been proposed as a solution to address the issue of high-quality data scarcity in the training of large language models (LLMs)
Our work delves into these specific flaws associated with question-answer (Q-A) pairs, a prevalent type of synthetic data, and presents a method based on unlearning techniques to mitigate these flaws.
Our work has yielded key insights into the effective use of synthetic data, aiming to promote more robust and efficient LLM training.
arXiv Detail & Related papers (2024-06-18T08:38:59Z) - DETAIL: Task DEmonsTration Attribution for Interpretable In-context Learning [75.68193159293425]
In-context learning (ICL) allows transformer-based language models to learn a specific task with a few "task demonstrations" without updating their parameters.
We propose an influence function-based attribution technique, DETAIL, that addresses the specific characteristics of ICL.
We experimentally prove the wide applicability of DETAIL by showing our attribution scores obtained on white-box models are transferable to black-box models in improving model performance.
arXiv Detail & Related papers (2024-05-22T15:52:52Z) - Learning Objective-Specific Active Learning Strategies with Attentive
Neural Processes [72.75421975804132]
Learning Active Learning (LAL) suggests to learn the active learning strategy itself, allowing it to adapt to the given setting.
We propose a novel LAL method for classification that exploits symmetry and independence properties of the active learning problem.
Our approach is based on learning from a myopic oracle, which gives our model the ability to adapt to non-standard objectives.
arXiv Detail & Related papers (2023-09-11T14:16:37Z) - Spatial-temporal Transformer for Affective Behavior Analysis [11.10521339384583]
We propose a Transformer with Multi-Head Attention framework to learn the distribution of both the spatial and temporal features.
The results fully demonstrate the effectiveness of our proposed model based on the Aff-Wild2 dataset.
arXiv Detail & Related papers (2023-03-19T04:34:17Z) - Facial Affect Analysis: Learning from Synthetic Data & Multi-Task
Learning Challenges [23.649517834303502]
We present novel frameworks to handle the two challenges in the 4th Affective Behavior Analysis In-The-Wild (ABAW) competition.
For MTL challenge, we adopt the SMM-EmotionNet with a better ensemble strategy of feature vectors.
For LSD challenge, we propose respective methods to combat the problems of single labels, imbalanced distribution, fine-tuning limitations, and choice of model architectures.
arXiv Detail & Related papers (2022-07-20T08:46:18Z) - HSE-NN Team at the 4th ABAW Competition: Multi-task Emotion Recognition
and Learning from Synthetic Images [7.056222499095849]
We present the results of the HSE-NN team in the 4th competition on Affective Behavior Analysis in-the-wild (ABAW)
The novel multi-task EfficientNet model is trained for simultaneous recognition of facial expressions.
The resulting MT-EmotiEffNet extracts visual features that are fed into simple feed-forward neural networks.
arXiv Detail & Related papers (2022-07-19T18:43:14Z) - Conformer-based Hybrid ASR System for Switchboard Dataset [99.88988282353206]
We present and evaluate a competitive conformer-based hybrid model training recipe.
We study different training aspects and methods to improve word-error-rate as well as to increase training speed.
We conduct experiments on Switchboard 300h dataset and our conformer-based hybrid model achieves competitive results.
arXiv Detail & Related papers (2021-11-05T12:03:18Z) - A Multi-task Mean Teacher for Semi-supervised Facial Affective Behavior
Analysis [15.95010869939508]
Existing affective behavior analysis method such as TSAV suffer from challenge of incomplete labeled datasets.
This paper presents a multi-task mean teacher model for semi-supervised Affective Behavior Analysis to learn from missing labels.
Experimental results on validation datasets show that our method achieves better performance than TSAV model.
arXiv Detail & Related papers (2021-07-09T05:48:22Z) - Task-Feature Collaborative Learning with Application to Personalized
Attribute Prediction [166.87111665908333]
We propose a novel multi-task learning method called Task-Feature Collaborative Learning (TFCL)
Specifically, we first propose a base model with a heterogeneous block-diagonal structure regularizer to leverage the collaborative grouping of features and tasks.
As a practical extension, we extend the base model by allowing overlapping features and differentiating the hard tasks.
arXiv Detail & Related papers (2020-04-29T02:32:04Z) - Gradient-Based Adversarial Training on Transformer Networks for
Detecting Check-Worthy Factual Claims [3.7543966923106438]
We introduce the first adversarially-regularized, transformer-based claim spotter model.
We obtain a 4.70 point F1-score improvement over current state-of-the-art models.
We propose a method to apply adversarial training to transformer models.
arXiv Detail & Related papers (2020-02-18T16:51:05Z) - Latent Opinions Transfer Network for Target-Oriented Opinion Words
Extraction [63.70885228396077]
We propose a novel model to transfer opinions knowledge from resource-rich review sentiment classification datasets to low-resource task TOWE.
Our model achieves better performance compared to other state-of-the-art methods and significantly outperforms the base model without transferring opinions knowledge.
arXiv Detail & Related papers (2020-01-07T11:50:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.