Meta Auxiliary Learning for Facial Action Unit Detection
- URL: http://arxiv.org/abs/2105.06620v1
- Date: Fri, 14 May 2021 02:28:40 GMT
- Title: Meta Auxiliary Learning for Facial Action Unit Detection
- Authors: Yong Li, Shiguang Shan
- Abstract summary: We consider learning AU detection and facial expression recognition in a multi-task manner.
The performance of the AU detection task cannot be always enhanced due to the negative transfer in the multi-task scenario.
We propose a Meta Learning method (MAL) that automatically selects highly related FE samples by learning adaptative weights for the training FE samples in a meta learning manner.
- Score: 84.22521265124806
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the success of deep neural networks on facial action unit (AU)
detection, better performance depends on a large number of training images with
accurate AU annotations. However, labeling AU is time-consuming, expensive, and
error-prone. Considering AU detection and facial expression recognition (FER)
are two highly correlated tasks, and facial expression (FE) is relatively easy
to annotate, we consider learning AU detection and FER in a multi-task manner.
However, the performance of the AU detection task cannot be always enhanced due
to the negative transfer in the multi-task scenario. To alleviate this issue,
we propose a Meta Auxiliary Learning method (MAL) that automatically selects
highly related FE samples by learning adaptative weights for the training FE
samples in a meta learning manner. The learned sample weights alleviate the
negative transfer from two aspects: 1) balance the loss of each task
automatically, and 2) suppress the weights of FE samples that have large
uncertainties. Experimental results on several popular AU datasets demonstrate
MAL consistently improves the AU detection performance compared with the
state-of-the-art multi-task and auxiliary learning methods. MAL automatically
estimates adaptive weights for the auxiliary FE samples according to their
semantic relevance with the primary AU detection task.
Related papers
- Facial Action Unit Detection by Adaptively Constraining Self-Attention and Causally Deconfounding Sample [53.23474626420103]
Facial action unit (AU) detection remains a challenging task, due to the subtlety, dynamics, and diversity of AUs.
We propose a novel AU detection framework called AC2D by adaptively constraining self-attention weight distribution.
Our method achieves competitive performance compared to state-of-the-art AU detection approaches on challenging benchmarks.
arXiv Detail & Related papers (2024-10-02T05:51:24Z) - Efficient Meta-Learning Enabled Lightweight Multiscale Few-Shot Object Detection in Remote Sensing Images [15.12889076965307]
YOLOv7 one-stage detector is subjected to a novel meta-learning training framework.
This transformation allows the detector to adeptly address FSOD tasks while capitalizing on its inherent advantage of lightweight.
To validate the effectiveness of our proposed detector, we conducted performance comparisons with current state-of-the-art detectors.
arXiv Detail & Related papers (2024-04-29T04:56:52Z) - Learning Contrastive Feature Representations for Facial Action Unit Detection [13.834540490373818]
Facial action unit (AU) detection has long encountered the challenge of detecting subtle feature differences when AUs activate.
We introduce a novel contrastive learning framework aimed for AU detection that incorporates both self-supervised and supervised signals.
arXiv Detail & Related papers (2024-02-09T03:48:20Z) - Multi-scale Promoted Self-adjusting Correlation Learning for Facial
Action Unit Detection [37.841035367349434]
Facial Action Unit (AU) detection is a crucial task in affective computing and social robotics.
Previous methods used fixed AU correlations based on expert experience or statistical rules on specific benchmarks.
This paper proposes a novel self-adjusting AU-correlation learning (SACL) method with less proposes for AU detection.
arXiv Detail & Related papers (2023-08-15T13:43:48Z) - DETA: Denoised Task Adaptation for Few-Shot Learning [135.96805271128645]
Test-time task adaptation in few-shot learning aims to adapt a pre-trained task-agnostic model for capturing taskspecific knowledge.
With only a handful of samples available, the adverse effect of either the image noise (a.k.a. X-noise) or the label noise (a.k.a. Y-noise) from support samples can be severely amplified.
We propose DEnoised Task Adaptation (DETA), a first, unified image- and label-denoising framework to existing task adaptation approaches.
arXiv Detail & Related papers (2023-03-11T05:23:20Z) - Towards Discriminative and Transferable One-Stage Few-Shot Object
Detectors [3.9189402702217344]
Few-shot object detection (FSOD) aims to address this problem by learning novel classes given only a few samples.
We make the observation that the large gap in performance between two-stage and one-stage FSODs are mainly due to their weak discriminability.
To address these limitations, we propose the Few-shot RetinaNet (FSRN) that consists of: a multi-way support training strategy to augment the number of foreground samples for dense meta-detectors.
arXiv Detail & Related papers (2022-10-11T20:58:25Z) - Cross-subject Action Unit Detection with Meta Learning and
Transformer-based Relation Modeling [7.395396464857193]
The paper proposes a meta-learning-based cross-subject AU detection model to eliminate the identity-caused differences.
A transformer-based relation learning module is introduced to learn the latent relations of multiple AUs.
Our results prove that on the two public datasets BP4D and DISFA, our method is superior to the state-of-the-art technology.
arXiv Detail & Related papers (2022-05-18T08:17:59Z) - Multi-Scale Positive Sample Refinement for Few-Shot Object Detection [61.60255654558682]
Few-shot object detection (FSOD) helps detectors adapt to unseen classes with few training instances.
We propose a Multi-scale Positive Sample Refinement (MPSR) approach to enrich object scales in FSOD.
MPSR generates multi-scale positive samples as object pyramids and refines the prediction at various scales.
arXiv Detail & Related papers (2020-07-18T09:48:29Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z) - Deep Multi-task Multi-label CNN for Effective Facial Attribute
Classification [53.58763562421771]
We propose a novel deep multi-task multi-label CNN, termed DMM-CNN, for effective Facial Attribute Classification (FAC)
Specifically, DMM-CNN jointly optimize two closely-related tasks (i.e., facial landmark detection and FAC) to improve the performance of FAC by taking advantage of multi-task learning.
Two different network architectures are respectively designed to extract features for two groups of attributes, and a novel dynamic weighting scheme is proposed to automatically assign the loss weight to each facial attribute during training.
arXiv Detail & Related papers (2020-02-10T12:34:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.