Self-Supervised Robustifying Guidance for Monocular 3D Face
Reconstruction
- URL: http://arxiv.org/abs/2112.14382v1
- Date: Wed, 29 Dec 2021 03:30:50 GMT
- Title: Self-Supervised Robustifying Guidance for Monocular 3D Face
Reconstruction
- Authors: Hitika Tiwari, Min-Hung Chen, Yi-Min Tsai, Hsien-Kai Kuo, Hung-Jen
Chen, Kevin Jou, K. S. Venkatesh, Yong-Sheng Chen
- Abstract summary: We propose a Self-Supervised RObustifying GUidancE (ROGUE) framework to obtain robustness against occlusions and noise in the face images.
The proposed network contains 1) the Guidance Pipeline to obtain the 3D face coefficients for the clean faces, and 2) the Robustification Pipeline to acquire the consistency between the estimated coefficients for occluded or noisy images and the clean counterpart.
On the three variations of the test dataset of CelebA, our method outperforms the current state-of-the-art method by large margins.
- Score: 14.203990541030445
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite the recent developments in 3D Face Reconstruction from occluded and
noisy face images, the performance is still unsatisfactory. One of the main
challenges is to handle moderate to heavy occlusions in the face images. In
addition, the noise in the face images inhibits the correct capture of facial
attributes, thus needing to be reliably addressed. Moreover, most existing
methods rely on additional dependencies, posing numerous constraints over the
training procedure. Therefore, we propose a Self-Supervised RObustifying
GUidancE (ROGUE) framework to obtain robustness against occlusions and noise in
the face images. The proposed network contains 1) the Guidance Pipeline to
obtain the 3D face coefficients for the clean faces, and 2) the Robustification
Pipeline to acquire the consistency between the estimated coefficients for
occluded or noisy images and the clean counterpart. The proposed image- and
feature-level loss functions aid the ROGUE learning process without posing
additional dependencies. On the three variations of the test dataset of CelebA:
rational occlusions, delusional occlusions, and noisy face images, our method
outperforms the current state-of-the-art method by large margins (e.g., for the
shape-based 3D vertex errors, a reduction from 0.146 to 0.048 for rational
occlusions, from 0.292 to 0.061 for delusional occlusions and from 0.269 to
0.053 for the noise in the face images), demonstrating the effectiveness of the
proposed approach.
Related papers
- 3D Priors-Guided Diffusion for Blind Face Restoration [30.94188504133298]
Blind face restoration endeavors to restore a clear face image from a degraded counterpart.
Recent approaches employing Generative Adversarial Networks (GANs) as priors have demonstrated remarkable success.
We propose a novel diffusion-based framework by embedding the 3D facial priors as structure and identity constraints into a denoising diffusion process.
arXiv Detail & Related papers (2024-09-02T07:13:32Z) - DPMesh: Exploiting Diffusion Prior for Occluded Human Mesh Recovery [71.6345505427213]
DPMesh is an innovative framework for occluded human mesh recovery.
It capitalizes on the profound diffusion prior about object structure and spatial relationships embedded in a pre-trained text-to-image diffusion model.
arXiv Detail & Related papers (2024-04-01T18:59:13Z) - I2V: Towards Texture-Aware Self-Supervised Blind Denoising using
Self-Residual Learning for Real-World Images [8.763680382529412]
pixel-shuffle downsampling (PD) has been proposed to eliminate the spatial correlation of noise.
We propose self-residual learning without the PD process to maintain texture information.
The results of extensive experiments show that the proposed method outperforms state-of-the-art self-supervised blind denoising approaches.
arXiv Detail & Related papers (2023-02-21T08:51:17Z) - Semantic-aware One-shot Face Re-enactment with Dense Correspondence
Estimation [100.60938767993088]
One-shot face re-enactment is a challenging task due to the identity mismatch between source and driving faces.
This paper proposes to use 3D Morphable Model (3DMM) for explicit facial semantic decomposition and identity disentanglement.
arXiv Detail & Related papers (2022-11-23T03:02:34Z) - Non-Deterministic Face Mask Removal Based On 3D Priors [3.8502825594372703]
The proposed approach integrates a multi-task 3D face reconstruction module with a face inpainting module.
By gradually controlling the 3D shape parameters, our method generates high-quality dynamic inpainting results with different expressions and mouth movements.
arXiv Detail & Related papers (2022-02-20T16:27:44Z) - Segmentation-Reconstruction-Guided Facial Image De-occlusion [48.952656891182826]
Occlusions are very common in face images in the wild, leading to the degraded performance of face-related tasks.
This paper proposes a novel face de-occlusion model based on face segmentation and 3D face reconstruction.
arXiv Detail & Related papers (2021-12-15T10:40:08Z) - Locality-aware Channel-wise Dropout for Occluded Face Recognition [116.2355331029041]
Face recognition is a challenging task in unconstrained scenarios, especially when faces are partially occluded.
We propose a novel and elegant occlusion-simulation method via dropping the activations of a group of neurons in some elaborately selected channel.
Experiments on various benchmarks show that the proposed method outperforms state-of-the-art methods with a remarkable improvement.
arXiv Detail & Related papers (2021-07-20T05:53:14Z) - Learning to Aggregate and Personalize 3D Face from In-the-Wild Photo
Collection [65.92058628082322]
Non-parametric face modeling aims to reconstruct 3D face only from images without shape assumptions.
This paper presents a novel Learning to Aggregate and Personalize framework for unsupervised robust 3D face modeling.
arXiv Detail & Related papers (2021-06-15T03:10:17Z) - Image Super-Resolution via Iterative Refinement [53.57766722279425]
SR3 is an approach to image Super-Resolution via Repeated Refinement.
It adapts probabilistic denoising diffusion models to conditional image generation.
It exhibits strong performance on super-resolution tasks at different magnification factors.
arXiv Detail & Related papers (2021-04-15T17:50:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.