Lossless Attention in Convolutional Networks for Facial Expression
Recognition in the Wild
- URL: http://arxiv.org/abs/2001.11869v1
- Date: Fri, 31 Jan 2020 14:38:35 GMT
- Title: Lossless Attention in Convolutional Networks for Facial Expression
Recognition in the Wild
- Authors: Chuang Wang, Ruimin Hu, Min Hu, Jiang Liu, Ting Ren, Shan He, Ming
Jiang, Jing Miao
- Abstract summary: We propose a Lossless Attention Model (LLAM) for convolutional neural networks (CNN) to extract attention-aware features from faces.
We participate in the seven basic expression classification sub-challenges of FG-2020 Affective Behavior Analysis in-the-wild Challenge.
And we validate our method on the Aff-Wild2 datasets released by the Challenge.
- Score: 26.10189921938026
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unlike the constraint frontal face condition, faces in the wild have various
unconstrained interference factors, such as complex illumination, changing
perspective and various occlusions. Facial expressions recognition (FER) in the
wild is a challenging task and existing methods can't perform well. However,
for occluded faces (containing occlusion caused by other objects and
self-occlusion caused by head posture changes), the attention mechanism has the
ability to focus on the non-occluded regions automatically. In this paper, we
propose a Lossless Attention Model (LLAM) for convolutional neural networks
(CNN) to extract attention-aware features from faces. Our module avoids decay
information in the process of generating attention maps by using the
information of the previous layer and not reducing the dimensionality.
Sequentially, we adaptively refine the feature responses by fusing the
attention map with the feature map. We participate in the seven basic
expression classification sub-challenges of FG-2020 Affective Behavior Analysis
in-the-wild Challenge. And we validate our method on the Aff-Wild2 datasets
released by the Challenge. The total accuracy (Accuracy) and the unweighted
mean (F1) of our method on the validation set are 0.49 and 0.38 respectively,
and the final result is 0.42 (0.67 F1-Score + 0.33 Accuracy).
Related papers
- Feature Attenuation of Defective Representation Can Resolve Incomplete Masking on Anomaly Detection [1.0358639819750703]
In unsupervised anomaly detection (UAD) research, it is necessary to develop a computationally efficient and scalable solution.
We revisit the reconstruction-by-inpainting approach and rethink to improve it by analyzing strengths and weaknesses.
We propose Feature Attenuation of Defective Representation (FADeR) that only employs two layers which attenuates feature information of anomaly reconstruction.
arXiv Detail & Related papers (2024-07-05T15:44:53Z) - Latent-OFER: Detect, Mask, and Reconstruct with Latent Vectors for
Occluded Facial Expression Recognition [0.0]
The proposed method can detect occluded parts of the face as if they were unoccluded, and recognize them, improving FER accuracy.
It involves three steps: First, the vision transformer (ViT)-based occlusion patch detector masks the occluded position by training only latent vectors from the unoccluded patches.
Second, the hybrid reconstruction network generates the masking position as a complete image using the ViT and convolutional neural network (CNN)
Last, the expression-relevant latent vector extractor retrieves and uses expression-related information from all latent vectors by applying a CNN-based class activation map
arXiv Detail & Related papers (2023-07-21T07:56:32Z) - Set-Based Face Recognition Beyond Disentanglement: Burstiness
Suppression With Variance Vocabulary [78.203301910422]
We argue that the two crucial issues in SFR, the face quality and burstiness, are both identity-irrelevant and variance-relevant.
We propose a light-weighted set-based disentanglement framework to separate the identity features with the variance features.
To suppress face burstiness in the sets, we propose a vocabulary-based burst suppression (VBS) method.
arXiv Detail & Related papers (2023-04-13T04:02:58Z) - Attribute-preserving Face Dataset Anonymization via Latent Code
Optimization [64.4569739006591]
We present a task-agnostic anonymization procedure that directly optimize the images' latent representation in the latent space of a pre-trained GAN.
We demonstrate through a series of experiments that our method is capable of anonymizing the identity of the images whilst -- crucially -- better-preserving the facial attributes.
arXiv Detail & Related papers (2023-03-20T17:34:05Z) - Robust and Precise Facial Landmark Detection by Self-Calibrated Pose
Attention Network [73.56802915291917]
We propose a semi-supervised framework to achieve more robust and precise facial landmark detection.
A Boundary-Aware Landmark Intensity (BALI) field is proposed to model more effective facial shape constraints.
A Self-Calibrated Pose Attention (SCPA) model is designed to provide a self-learned objective function that enforces intermediate supervision.
arXiv Detail & Related papers (2021-12-23T02:51:08Z) - End2End Occluded Face Recognition by Masking Corrupted Features [82.27588990277192]
State-of-the-art general face recognition models do not generalize well to occluded face images.
This paper presents a novel face recognition method that is robust to occlusions based on a single end-to-end deep neural network.
Our approach, named FROM (Face Recognition with Occlusion Masks), learns to discover the corrupted features from the deep convolutional neural networks, and clean them by the dynamically learned masks.
arXiv Detail & Related papers (2021-08-21T09:08:41Z) - Robust Facial Expression Recognition with Convolutional Visual
Transformers [23.05378099875569]
We propose Convolutional Visual Transformers to tackle Facial Expression Recognition in the wild by two main steps.
First, we propose an attentional selective fusion (ASF) for leveraging the feature maps generated by two-branch CNNs.
Second, inspired by the success of Transformers in natural language processing, we propose to model relationships between these visual words with global self-attention.
arXiv Detail & Related papers (2021-03-31T07:07:56Z) - Micro-Facial Expression Recognition Based on Deep-Rooted Learning
Algorithm [0.0]
An effective Micro-Facial Expression Based Deep-Rooted Learning (MFEDRL) classifier is proposed in this paper.
The performance of the algorithm will be evaluated using recognition rate and false measures.
arXiv Detail & Related papers (2020-09-12T12:23:27Z) - Occlusion-Adaptive Deep Network for Robust Facial Expression Recognition [56.11054589916299]
We propose a landmark-guided attention branch to find and discard corrupted features from occluded regions.
An attention map is first generated to indicate if a specific facial part is occluded and guide our model to attend to non-occluded regions.
This results in more diverse and discriminative features, enabling the expression recognition system to recover even though the face is partially occluded.
arXiv Detail & Related papers (2020-05-12T20:42:55Z) - Suppressing Uncertainties for Large-Scale Facial Expression Recognition [81.51495681011404]
This paper proposes a simple yet efficient Self-Cure Network (SCN) which suppresses the uncertainties efficiently and prevents deep networks from over-fitting uncertain facial images.
Results on public benchmarks demonstrate that our SCN outperforms current state-of-the-art methods with textbf88.14% on RAF-DB, textbf60.23% on AffectNet, and textbf89.35% on FERPlus.
arXiv Detail & Related papers (2020-02-24T17:24:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.