Face Frontalization Based on Robustly Fitting a Deformable Shape Model
to 3D Landmarks
- URL: http://arxiv.org/abs/2010.13676v2
- Date: Wed, 10 Mar 2021 10:45:41 GMT
- Title: Face Frontalization Based on Robustly Fitting a Deformable Shape Model
to 3D Landmarks
- Authors: Zhiqi Kang, Mostafa Sadeghi and Radu Horaud
- Abstract summary: Face frontalization consists of a frontally-viewed face from an arbitrarily-viewed one.
The main contribution of this paper is a robust face alignment method that enables pixel-to-pixel warping.
An important merit of the proposed method is its ability to deal both with noise (small perturbations) and with outliers (large errors)
- Score: 24.07648367866321
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Face frontalization consists of synthesizing a frontally-viewed face from an
arbitrarily-viewed one. The main contribution of this paper is a robust face
alignment method that enables pixel-to-pixel warping. The method simultaneously
estimates the rigid transformation (scale, rotation, and translation) and the
non-rigid deformation between two 3D point sets: a set of 3D landmarks
extracted from an arbitrary-viewed face, and a set of 3D landmarks
parameterized by a frontally-viewed deformable face model. An important merit
of the proposed method is its ability to deal both with noise (small
perturbations) and with outliers (large errors). We propose to model inliers
and outliers with the generalized Student's t-probability distribution
function, a heavy-tailed distribution that is immune to non-Gaussian errors in
the data. We describe in detail the associated expectation-maximization (EM)
algorithm that alternates between the estimation of (i) the rigid parameters,
(ii) the deformation parameters, and (iii) the Student-t distribution
parameters. We also propose to use the zero-mean normalized cross-correlation,
between a frontalized face and the corresponding ground-truth frontally-viewed
face, to evaluate the performance of frontalization. To this end, we use a
dataset that contains pairs of profile-viewed and frontally-viewed faces. This
evaluation, based on direct image-to-image comparison, stands in contrast with
indirect evaluation, based on analyzing the effect of frontalization on face
recognition.
Related papers
- OFER: Occluded Face Expression Reconstruction [16.06622406877353]
We introduce OFER, a novel approach for single image 3D face reconstruction that can generate plausible, diverse, and expressive 3D faces.
We propose a novel ranking mechanism that sorts the outputs of the shape diffusion network based on the predicted shape accuracy scores to select the best match.
arXiv Detail & Related papers (2024-10-29T00:21:26Z) - Semantic-aware One-shot Face Re-enactment with Dense Correspondence
Estimation [100.60938767993088]
One-shot face re-enactment is a challenging task due to the identity mismatch between source and driving faces.
This paper proposes to use 3D Morphable Model (3DMM) for explicit facial semantic decomposition and identity disentanglement.
arXiv Detail & Related papers (2022-11-23T03:02:34Z) - Information Maximization for Extreme Pose Face Recognition [18.545778273606427]
We exploit a connection using a coupled-encoder network to project frontal/profile face images into a common embedding space.
The proposed model forces the similarity of representations in the embedding space by maximizing the mutual information between two views of the face.
arXiv Detail & Related papers (2022-09-07T20:30:06Z) - Expression-preserving face frontalization improves visually assisted
speech processing [35.647888055229956]
The main contribution of this paper is a frontalization methodology that preserves non-rigid facial deformations.
We show that the method, when incorporated into deep learning pipelines, improves word recognition and speech intelligibilty scores by a considerable margin.
arXiv Detail & Related papers (2022-04-06T13:22:24Z) - Uncertainty-Aware Adaptation for Self-Supervised 3D Human Pose
Estimation [70.32536356351706]
We introduce MRP-Net that constitutes a common deep network backbone with two output heads subscribing to two diverse configurations.
We derive suitable measures to quantify prediction uncertainty at both pose and joint level.
We present a comprehensive evaluation of the proposed approach and demonstrate state-of-the-art performance on benchmark datasets.
arXiv Detail & Related papers (2022-03-29T07:14:58Z) - A Model for Multi-View Residual Covariances based on Perspective
Deformation [88.21738020902411]
We derive a model for the covariance of the visual residuals in multi-view SfM, odometry and SLAM setups.
We validate our model with synthetic and real data and integrate it into photometric and feature-based Bundle Adjustment.
arXiv Detail & Related papers (2022-02-01T21:21:56Z) - Adversarial Parametric Pose Prior [106.12437086990853]
We learn a prior that restricts the SMPL parameters to values that produce realistic poses via adversarial training.
We show that our learned prior covers the diversity of the real-data distribution, facilitates optimization for 3D reconstruction from 2D keypoints, and yields better pose estimates when used for regression from images.
arXiv Detail & Related papers (2021-12-08T10:05:32Z) - Learning to Aggregate and Personalize 3D Face from In-the-Wild Photo
Collection [65.92058628082322]
Non-parametric face modeling aims to reconstruct 3D face only from images without shape assumptions.
This paper presents a novel Learning to Aggregate and Personalize framework for unsupervised robust 3D face modeling.
arXiv Detail & Related papers (2021-06-15T03:10:17Z) - Implicit-PDF: Non-Parametric Representation of Probability Distributions
on the Rotation Manifold [47.31074799708132]
We introduce a method to estimate arbitrary, non-parametric distributions on SO(3).
Our key idea is to represent the distributions implicitly, with a neural network that estimates the probability given the input image and a candidate pose.
We achieve state-of-the-art performance on Pascal3D+ and ModelNet10-SO(3) benchmarks.
arXiv Detail & Related papers (2021-06-10T17:57:23Z) - Multi-person 3D Pose Estimation in Crowded Scenes Based on Multi-View
Geometry [62.29762409558553]
Epipolar constraints are at the core of feature matching and depth estimation in multi-person 3D human pose estimation methods.
Despite the satisfactory performance of this formulation in sparser crowd scenes, its effectiveness is frequently challenged under denser crowd circumstances.
In this paper, we depart from the multi-person 3D pose estimation formulation, and instead reformulate it as crowd pose estimation.
arXiv Detail & Related papers (2020-07-21T17:59:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.