Facial UV Map Completion for Pose-invariant Face Recognition: A Novel
Adversarial Approach based on Coupled Attention Residual UNets
- URL: http://arxiv.org/abs/2011.00912v1
- Date: Mon, 2 Nov 2020 11:46:42 GMT
- Title: Facial UV Map Completion for Pose-invariant Face Recognition: A Novel
Adversarial Approach based on Coupled Attention Residual UNets
- Authors: In Seop Na, Chung Tran, Dung Nguyen and Sang Dinh
- Abstract summary: We propose a novel generative model called Attention ResCUNet-GAN to improve the UV map completion.
We show that the proposed method yields superior performance compared to other existing methods.
- Score: 3.999563862575646
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Pose-invariant face recognition refers to the problem of identifying or
verifying a person by analyzing face images captured from different poses. This
problem is challenging due to the large variation of pose, illumination and
facial expression. A promising approach to deal with pose variation is to
fulfill incomplete UV maps extracted from in-the-wild faces, then attach the
completed UV map to a fitted 3D mesh and finally generate different 2D faces of
arbitrary poses. The synthesized faces increase the pose variation for training
deep face recognition models and reduce the pose discrepancy during the testing
phase. In this paper, we propose a novel generative model called Attention
ResCUNet-GAN to improve the UV map completion. We enhance the original UV-GAN
by using a couple of U-Nets. Particularly, the skip connections within each
U-Net are boosted by attention gates. Meanwhile, the features from two U-Nets
are fused with trainable scalar weights. The experiments on the popular
benchmarks, including Multi-PIE, LFW, CPLWF and CFP datasets, show that the
proposed method yields superior performance compared to other existing methods.
Related papers
- Realistic and Efficient Face Swapping: A Unified Approach with Diffusion Models [69.50286698375386]
We propose a novel approach that better harnesses diffusion models for face-swapping.
We introduce a mask shuffling technique during inpainting training, which allows us to create a so-called universal model for swapping.
Ours is a relatively unified approach and so it is resilient to errors in other off-the-shelf models.
arXiv Detail & Related papers (2024-09-11T13:43:53Z) - UVMap-ID: A Controllable and Personalized UV Map Generative Model [67.71022515856653]
We introduce UVMap-ID, a controllable and personalized UV Map generative model.
Unlike traditional large-scale training methods in 2D, we propose to fine-tune a pre-trained text-to-image diffusion model.
Both quantitative and qualitative analyses demonstrate the effectiveness of our method in controllable and personalized UV Map generation.
arXiv Detail & Related papers (2024-04-22T20:30:45Z) - POCE: Pose-Controllable Expression Editing [75.7701103792032]
This paper presents POCE, an innovative pose-controllable expression editing network.
It can generate realistic facial expressions and head poses simultaneously with just unpaired training images.
The learned model can generate realistic and high-fidelity facial expressions under various new poses.
arXiv Detail & Related papers (2023-04-18T12:26:19Z) - Towards Hard-pose Virtual Try-on via 3D-aware Global Correspondence
Learning [70.75369367311897]
3D-aware global correspondences are reliable flows that jointly encode global semantic correlations, local deformations, and geometric priors of 3D human bodies.
An adversarial generator takes the garment warped by the 3D-aware flow, and the image of the target person as inputs, to synthesize the photo-realistic try-on result.
arXiv Detail & Related papers (2022-11-25T12:16:21Z) - FSGANv2: Improved Subject Agnostic Face Swapping and Reenactment [28.83743270895698]
We present Face Swapping GAN (FSGAN) for face swapping and reenactment.
Unlike previous work, we offer a subject swapping scheme that can be applied to pairs of faces without requiring training on those faces.
We derive a novel iterative deep learning--based approach for face reenactment which adjusts significant pose and expression variations that can be applied to a single image or a video sequence.
For video sequences, we introduce a continuous agnostic of the face views based on reenactment, Delaunay Triangulation, and bary coordinates. Occluded face regions are handled by a face completion network.
arXiv Detail & Related papers (2022-02-25T21:04:39Z) - Learning to Aggregate and Personalize 3D Face from In-the-Wild Photo
Collection [65.92058628082322]
Non-parametric face modeling aims to reconstruct 3D face only from images without shape assumptions.
This paper presents a novel Learning to Aggregate and Personalize framework for unsupervised robust 3D face modeling.
arXiv Detail & Related papers (2021-06-15T03:10:17Z) - Pose Invariant Person Re-Identification using Robust Pose-transformation
GAN [11.338815177557645]
Person re-identification (re-ID) aims to retrieve a person's images from an image gallery, given a single instance of the person of interest.
Despite several advancements, learning discriminative identity-sensitive and viewpoint invariant features for robust Person Re-identification is a major challenge owing to large pose variation of humans.
This paper proposes a re-ID pipeline that utilizes the image generation capability of Generative Adversarial Networks combined with pose regression and feature fusion to achieve pose invariant feature learning.
arXiv Detail & Related papers (2021-04-11T15:47:03Z) - SMAP: Single-Shot Multi-Person Absolute 3D Pose Estimation [46.85865451812981]
We propose a novel system that first regresses a set of 2.5D representations of body parts and then reconstructs the 3D absolute poses based on these 2.5D representations with a depth-aware part association algorithm.
Such a single-shot bottom-up scheme allows the system to better learn and reason about the inter-person depth relationship, improving both 3D and 2D pose estimation.
arXiv Detail & Related papers (2020-08-26T09:56:07Z) - Self-Supervised Monocular 3D Face Reconstruction by Occlusion-Aware
Multi-view Geometry Consistency [40.56510679634943]
We propose a self-supervised training architecture by leveraging the multi-view geometry consistency.
We design three novel loss functions for multi-view consistency, including the pixel consistency loss, the depth consistency loss, and the facial landmark-based epipolar loss.
Our method is accurate and robust, especially under large variations of expressions, poses, and illumination conditions.
arXiv Detail & Related papers (2020-07-24T12:36:09Z) - Fusing Wearable IMUs with Multi-View Images for Human Pose Estimation: A
Geometric Approach [76.10879433430466]
We propose to estimate 3D human pose from multi-view images and a few IMUs attached at person's limbs.
It operates by firstly detecting 2D poses from the two signals, and then lifting them to the 3D space.
The simple two-step approach reduces the error of the state-of-the-art by a large margin on a public dataset.
arXiv Detail & Related papers (2020-03-25T00:26:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.