Multitask Emotion Recognition Model with Knowledge Distillation and Task
Discriminator
- URL: http://arxiv.org/abs/2203.13072v1
- Date: Thu, 24 Mar 2022 13:50:48 GMT
- Title: Multitask Emotion Recognition Model with Knowledge Distillation and Task
Discriminator
- Authors: Euiseok Jeong, Geesung Oh and Sejoon Lim
- Abstract summary: We designed a multi-task model using ABAW dataset to predict emotions.
We trained model from the incomplete label by applying the knowledge distillation technique.
As a result we achieved 2.40 in Multi Task Learning task validation dataset.
- Score: 0.0
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Due to the collection of big data and the development of deep learning,
research to predict human emotions in the wild is being actively conducted. We
designed a multi-task model using ABAW dataset to predict valence-arousal,
expression, and action unit through audio data and face images at in real
world. We trained model from the incomplete label by applying the knowledge
distillation technique. The teacher model was trained as a supervised learning
method, and the student model was trained by using the output of the teacher
model as a soft label. As a result we achieved 2.40 in Multi Task Learning task
validation dataset.
Related papers
- Interactive DualChecker for Mitigating Hallucinations in Distilling Large Language Models [7.632217365130212]
Large Language Models (LLMs) have demonstrated exceptional capabilities across various machine learning (ML) tasks.
These models can produce hallucinations, particularly in domains with incomplete knowledge.
We introduce DualChecker, an innovative framework designed to mitigate hallucinations and improve the performance of both teacher and student models.
arXiv Detail & Related papers (2024-08-22T12:04:04Z) - Diffusion Model is an Effective Planner and Data Synthesizer for
Multi-Task Reinforcement Learning [101.66860222415512]
Multi-Task Diffusion Model (textscMTDiff) is a diffusion-based method that incorporates Transformer backbones and prompt learning for generative planning and data synthesis.
For generative planning, we find textscMTDiff outperforms state-of-the-art algorithms across 50 tasks on Meta-World and 8 maps on Maze2D.
arXiv Detail & Related papers (2023-05-29T05:20:38Z) - Ensemble knowledge distillation of self-supervised speech models [84.69577440755457]
Distilled self-supervised models have shown competitive performance and efficiency in recent years.
We performed Ensemble Knowledge Distillation (EKD) on various self-supervised speech models such as HuBERT, RobustHuBERT, and WavLM.
Our method improves the performance of the distilled models on four downstream speech processing tasks.
arXiv Detail & Related papers (2023-02-24T17:15:39Z) - Prototype-guided Cross-task Knowledge Distillation for Large-scale
Models [103.04711721343278]
Cross-task knowledge distillation helps to train a small student model to obtain a competitive performance.
We propose a Prototype-guided Cross-task Knowledge Distillation (ProC-KD) approach to transfer the intrinsic local-level object knowledge of a large-scale teacher network to various task scenarios.
arXiv Detail & Related papers (2022-12-26T15:00:42Z) - Multi-Task Self-Training for Learning General Representations [97.01728635294879]
Multi-task self-training (MuST) harnesses the knowledge in independent specialized teacher models to train a single general student model.
MuST is scalable with unlabeled or partially labeled datasets and outperforms both specialized supervised models and self-supervised models when training on large scale datasets.
arXiv Detail & Related papers (2021-08-25T17:20:50Z) - A Multi-task Mean Teacher for Semi-supervised Facial Affective Behavior
Analysis [15.95010869939508]
Existing affective behavior analysis method such as TSAV suffer from challenge of incomplete labeled datasets.
This paper presents a multi-task mean teacher model for semi-supervised Affective Behavior Analysis to learn from missing labels.
Experimental results on validation datasets show that our method achieves better performance than TSAV model.
arXiv Detail & Related papers (2021-07-09T05:48:22Z) - Multitask Multi-database Emotion Recognition [1.52292571922932]
We introduce our submission to the 2nd Affective Behavior Analysis in-the-wild (ABAW) 2021 competition.
We train a unified deep learning model on multi- databases to perform two tasks.
Experiment results show that the network have achieved promising results on the validation set of the AffWild2 database.
arXiv Detail & Related papers (2021-07-08T21:57:58Z) - Reinforced Multi-Teacher Selection for Knowledge Distillation [54.72886763796232]
knowledge distillation is a popular method for model compression.
Current methods assign a fixed weight to a teacher model in the whole distillation.
Most of the existing methods allocate an equal weight to every teacher model.
In this paper, we observe that, due to the complexity of training examples and the differences in student model capability, learning differentially from teacher models can lead to better performance of student models distilled.
arXiv Detail & Related papers (2020-12-11T08:56:39Z) - Label-Efficient Multi-Task Segmentation using Contrastive Learning [0.966840768820136]
We propose a multi-task segmentation model with a contrastive learning based subtask and compare its performance with other multi-task models.
We experimentally show that our proposed method outperforms other multi-task methods including the state-of-the-art fully supervised model when the amount of annotated data is limited.
arXiv Detail & Related papers (2020-09-23T14:12:17Z) - Learning to Reweight with Deep Interactions [104.68509759134878]
We propose an improved data reweighting algorithm, in which the student model provides its internal states to the teacher model.
Experiments on image classification with clean/noisy labels and neural machine translation empirically demonstrate that our algorithm makes significant improvement over previous methods.
arXiv Detail & Related papers (2020-07-09T09:06:31Z) - Multitask Emotion Recognition with Incomplete Labels [7.811459544911892]
We train a unified model to perform three tasks: facial action unit detection, expression classification, and valence-arousal estimation.
Most existing datasets do not contain labels for all three tasks.
We find that most of the student models outperform their teacher model on all the three tasks.
arXiv Detail & Related papers (2020-02-10T05:32:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.