Global-to-local Expression-aware Embeddings for Facial Action Unit
Detection
- URL: http://arxiv.org/abs/2210.15160v2
- Date: Fri, 28 Oct 2022 02:42:34 GMT
- Title: Global-to-local Expression-aware Embeddings for Facial Action Unit
Detection
- Authors: Rudong An, Wei Zhang, Hao Zeng, Wei Chen, Zhigang Deng, Yu Ding
- Abstract summary: We propose a novel fine-grained textslGlobal Expression representation to capture subtle and continuous facial movements.
It consists of an AU feature map extractor and a corresponding AU mask extractor.
Our method validly outperforms previous works and achieves state-of-the-art performances on widely-used face datasets.
- Score: 18.629509376315752
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Expressions and facial action units (AUs) are two levels of facial behavior
descriptors. Expression auxiliary information has been widely used to improve
the AU detection performance. However, most existing expression representations
can only describe pre-determined discrete categories (e.g., Angry, Disgust,
Happy, Sad, etc.) and cannot capture subtle expression transformations like
AUs. In this paper, we propose a novel fine-grained \textsl{Global Expression
representation Encoder} to capture subtle and continuous facial movements, to
promote AU detection. To obtain such a global expression representation, we
propose to train an expression embedding model on a large-scale expression
dataset according to global expression similarity. Moreover, considering the
local definition of AUs, it is essential to extract local AU features.
Therefore, we design a \textsl{Local AU Features Module} to generate local
facial features for each AU. Specifically, it consists of an AU feature map
extractor and a corresponding AU mask extractor. First, the two extractors
transform the global expression representation into AU feature maps and masks,
respectively. Then, AU feature maps and their corresponding AU masks are
multiplied to generate AU masked features focusing on local facial region.
Finally, the AU masked features are fed into an AU classifier for judging the
AU occurrence. Extensive experiment results demonstrate the superiority of our
proposed method. Our method validly outperforms previous works and achieves
state-of-the-art performances on widely-used face datasets, including BP4D,
DISFA, and BP4D+.
Related papers
- Spatial Action Unit Cues for Interpretable Deep Facial Expression Recognition [55.97779732051921]
State-of-the-art classifiers for facial expression recognition (FER) lack interpretability, an important feature for end-users.
A new learning strategy is proposed to explicitly incorporate AU cues into classifier training, allowing to train deep interpretable models.
Our new strategy is generic, and can be applied to any deep CNN- or transformer-based classifier without requiring any architectural change or significant additional training time.
arXiv Detail & Related papers (2024-10-01T10:42:55Z) - MaskInversion: Localized Embeddings via Optimization of Explainability Maps [49.50785637749757]
MaskInversion generates a context-aware embedding for a query image region specified by a mask at test time.
It can be used for a broad range of tasks, including open-vocabulary class retrieval, referring expression comprehension, as well as for localized captioning and image generation.
arXiv Detail & Related papers (2024-07-29T14:21:07Z) - Guided Interpretable Facial Expression Recognition via Spatial Action Unit Cues [55.97779732051921]
A new learning strategy is proposed to explicitly incorporate au cues into classifier training.
We show that our strategy can improve layer-wise interpretability without degrading classification performance.
arXiv Detail & Related papers (2024-02-01T02:13:49Z) - Facial Action Units Detection Aided by Global-Local Expression Embedding [36.78982474775454]
We develop a novel AU detection framework aided by the Global-Local facial Expressions Embedding, dubbed GLEE-Net.
Our GLEE-Net consists of three branches to extract identity-independent expression features for AU detection.
Our method significantly outperforms the state-of-the-art on the widely-used DISFA, BP4D and BP4D+ datasets.
arXiv Detail & Related papers (2022-10-25T02:35:32Z) - Dynamic Prototype Mask for Occluded Person Re-Identification [88.7782299372656]
Existing methods mainly address this issue by employing body clues provided by an extra network to distinguish the visible part.
We propose a novel Dynamic Prototype Mask (DPM) based on two self-evident prior knowledge.
Under this condition, the occluded representation could be well aligned in a selected subspace spontaneously.
arXiv Detail & Related papers (2022-07-19T03:31:13Z) - AU-Expression Knowledge Constrained Representation Learning for Facial
Expression Recognition [79.8779790682205]
We propose an AU-Expression Knowledge Constrained Representation Learning (AUE-CRL) framework to learn the AU representations without AU annotations and adaptively use representations to facilitate facial expression recognition.
We conduct experiments on the challenging uncontrolled datasets to demonstrate the superiority of the proposed framework over current state-of-the-art methods.
arXiv Detail & Related papers (2020-12-29T03:42:04Z) - AU-Guided Unsupervised Domain Adaptive Facial Expression Recognition [21.126514122636966]
This paper proposes an AU-guided unsupervised Domain Adaptive FER framework to relieve the annotation bias between different FER datasets.
To achieve domain-invariant compact features, we utilize an AU-guided triplet training which randomly collects anchor-positive-negative triplets on both domains with AUs.
arXiv Detail & Related papers (2020-12-18T07:17:30Z) - J$\hat{\text{A}}$A-Net: Joint Facial Action Unit Detection and Face
Alignment via Adaptive Attention [57.51255553918323]
We propose a novel end-to-end deep learning framework for joint AU detection and face alignment.
Our framework significantly outperforms the state-of-the-art AU detection methods on the challenging BP4D, DISFA, GFT and BP4D+ benchmarks.
arXiv Detail & Related papers (2020-03-18T12:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.