Unsupervised Facial Action Unit Intensity Estimation via Differentiable
Optimization
- URL: http://arxiv.org/abs/2004.05908v1
- Date: Mon, 13 Apr 2020 12:56:28 GMT
- Title: Unsupervised Facial Action Unit Intensity Estimation via Differentiable
Optimization
- Authors: Xinhui Song and Tianyang Shi and Tianjia Shao and Yi Yuan and Zunlei
Feng and Changjie Fan
- Abstract summary: We propose an unsupervised framework GE-Net for facial AU intensity estimation from a single image.
Our framework performs differentiable optimization, which iteratively updates the facial parameters to match the input image.
Experimental results demonstrate that our method can achieve state-of-the-art results compared with existing methods.
- Score: 45.07851622835555
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The automatic intensity estimation of facial action units (AUs) from a single
image plays a vital role in facial analysis systems. One big challenge for
data-driven AU intensity estimation is the lack of sufficient AU label data.
Due to the fact that AU annotation requires strong domain expertise, it is
expensive to construct an extensive database to learn deep models. The limited
number of labeled AUs as well as identity differences and pose variations
further increases the estimation difficulties. Considering all these
difficulties, we propose an unsupervised framework GE-Net for facial AU
intensity estimation from a single image, without requiring any annotated AU
data. Our framework performs differentiable optimization, which iteratively
updates the facial parameters (i.e., head pose, AU parameters and identity
parameters) to match the input image. GE-Net consists of two modules: a
generator and a feature extractor. The generator learns to "render" a face
image from a set of facial parameters in a differentiable way, and the feature
extractor extracts deep features for measuring the similarity of the rendered
image and input real image. After the two modules are trained and fixed, the
framework searches optimal facial parameters by minimizing the differences of
the extracted features between the rendered image and the input image.
Experimental results demonstrate that our method can achieve state-of-the-art
results compared with existing methods.
Related papers
- Multi-Feature Aggregation in Diffusion Models for Enhanced Face Super-Resolution [6.055006354743854]
We develop an algorithm that utilize a low-resolution image combined with features extracted from multiple low-quality images to generate a super-resolved image.
Unlike other algorithms, our approach recovers facial features without explicitly providing attribute information.
This is the first time multi-features combined with low-resolution images are used as conditioners to generate more reliable super-resolution images.
arXiv Detail & Related papers (2024-08-27T20:08:33Z) - OCR is All you need: Importing Multi-Modality into Image-based Defect Detection System [7.1083241462091165]
We introduce an external modality-guided data mining framework, primarily rooted in optical character recognition (OCR), to extract statistical features from images.
A key aspect of our approach is the alignment of external modality features, extracted using a single modality-aware model, with image features encoded by a convolutional neural network.
Our methodology considerably boosts the recall rate of the defect detection model and maintains high robustness even in challenging scenarios.
arXiv Detail & Related papers (2024-03-18T07:41:39Z) - From Static to Dynamic: Adapting Landmark-Aware Image Models for Facial Expression Recognition in Videos [88.08209394979178]
Dynamic facial expression recognition (DFER) in the wild is still hindered by data limitations.
We introduce a novel Static-to-Dynamic model (S2D) that leverages existing SFER knowledge and dynamic information implicitly encoded in extracted facial landmark-aware features.
arXiv Detail & Related papers (2023-12-09T03:16:09Z) - DeepFidelity: Perceptual Forgery Fidelity Assessment for Deepfake
Detection [67.3143177137102]
Deepfake detection refers to detecting artificially generated or edited faces in images or videos.
We propose a novel Deepfake detection framework named DeepFidelity to adaptively distinguish real and fake faces.
arXiv Detail & Related papers (2023-12-07T07:19:45Z) - Dynamic Prototype Mask for Occluded Person Re-Identification [88.7782299372656]
Existing methods mainly address this issue by employing body clues provided by an extra network to distinguish the visible part.
We propose a novel Dynamic Prototype Mask (DPM) based on two self-evident prior knowledge.
Under this condition, the occluded representation could be well aligned in a selected subspace spontaneously.
arXiv Detail & Related papers (2022-07-19T03:31:13Z) - MorphGAN: One-Shot Face Synthesis GAN for Detecting Recognition Bias [13.162012586770576]
We describe a simulator that applies specific head pose and facial expression adjustments to images of previously unseen people.
We show that by augmenting small datasets of faces with new poses and expressions improves the recognition performance by up to 9% depending on the augmentation and data scarcity.
arXiv Detail & Related papers (2020-12-09T18:43:03Z) - Unsupervised Learning Facial Parameter Regressor for Action Unit
Intensity Estimation via Differentiable Renderer [51.926868759681014]
We present a framework to predict the facial parameters based on a bone-driven face model (BDFM) under different views.
The proposed framework consists of a feature extractor, a generator, and a facial parameter regressor.
arXiv Detail & Related papers (2020-08-20T09:49:13Z) - Multi-Scale Thermal to Visible Face Verification via Attribute Guided
Synthesis [55.29770222566124]
We use attributes extracted from visible images to synthesize attribute-preserved visible images from thermal imagery for cross-modal matching.
A novel multi-scale generator is proposed to synthesize the visible image from the thermal image guided by the extracted attributes.
A pre-trained VGG-Face network is leveraged to extract features from the synthesized image and the input visible image for verification.
arXiv Detail & Related papers (2020-04-20T01:45:05Z) - Vec2Face: Unveil Human Faces from their Blackbox Features in Face
Recognition [29.106746021757168]
This paper presents a novel generative structure with Bijective Metric Learning, namely Bijective Generative Adrial Networks in a Distillation framework (DiBiGAN)
Results on several benchmarking datasets have demonstrated the effectiveness of DiBiGAN on both image realism and ID preservation properties.
arXiv Detail & Related papers (2020-03-16T00:30:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.