Progressive Semantic-Aware Style Transformation for Blind Face
Restoration
- URL: http://arxiv.org/abs/2009.08709v2
- Date: Sun, 21 Mar 2021 09:35:05 GMT
- Title: Progressive Semantic-Aware Style Transformation for Blind Face
Restoration
- Authors: Chaofeng Chen, Xiaoming Li, Lingbo Yang, Xianhui Lin, Lei Zhang,
Kwan-Yee K. Wong
- Abstract summary: We propose a new progressive semantic-aware style transformation framework, named PSFR-GAN, for face restoration.
The proposed PSFR-GAN makes full use of the semantic (parsing maps) and pixel (LQ images) space information from different scales of input pairs.
Experiment results show that our model trained with synthetic data can not only produce more realistic high-resolution results for synthetic LQ inputs but also better to generalize natural LQ face images.
- Score: 26.66332852514812
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Face restoration is important in face image processing, and has been widely
studied in recent years. However, previous works often fail to generate
plausible high quality (HQ) results for real-world low quality (LQ) face
images. In this paper, we propose a new progressive semantic-aware style
transformation framework, named PSFR-GAN, for face restoration. Specifically,
instead of using an encoder-decoder framework as previous methods, we formulate
the restoration of LQ face images as a multi-scale progressive restoration
procedure through semantic-aware style transformation. Given a pair of LQ face
image and its corresponding parsing map, we first generate a multi-scale
pyramid of the inputs, and then progressively modulate different scale features
from coarse-to-fine in a semantic-aware style transfer way. Compared with
previous networks, the proposed PSFR-GAN makes full use of the semantic
(parsing maps) and pixel (LQ images) space information from different scales of
input pairs. In addition, we further introduce a semantic aware style loss
which calculates the feature style loss for each semantic region individually
to improve the details of face textures. Finally, we pretrain a face parsing
network which can generate decent parsing maps from real-world LQ face images.
Experiment results show that our model trained with synthetic data can not only
produce more realistic high-resolution results for synthetic LQ inputs and but
also generalize better to natural LQ face images compared with state-of-the-art
methods. Codes are available at https://github.com/chaofengc/PSFRGAN.
Related papers
- 3D Facial Expressions through Analysis-by-Neural-Synthesis [30.2749903946587]
SMIRK (Spatial Modeling for Image-based Reconstruction of Kinesics) faithfully reconstructs expressive 3D faces from images.
We identify two key limitations in existing methods: shortcomings in their self-supervised training formulation, and a lack of expression diversity in the training images.
Our qualitative, quantitative and particularly our perceptual evaluations demonstrate that SMIRK achieves the new state-of-the art performance on accurate expression reconstruction.
arXiv Detail & Related papers (2024-04-05T14:00:07Z) - CLR-Face: Conditional Latent Refinement for Blind Face Restoration Using
Score-Based Diffusion Models [57.9771859175664]
Recent generative-prior-based methods have shown promising blind face restoration performance.
Generating fine-grained facial details faithful to inputs remains a challenging problem.
We introduce a diffusion-based-prior inside a VQGAN architecture that focuses on learning the distribution over uncorrupted latent embeddings.
arXiv Detail & Related papers (2024-02-08T23:51:49Z) - SARGAN: Spatial Attention-based Residuals for Facial Expression
Manipulation [1.7056768055368383]
We present a novel method named SARGAN that addresses the limitations from three perspectives.
We exploited a symmetric encoder-decoder network to attend facial features at multiple scales.
Our proposed model performs significantly better than state-of-the-art methods.
arXiv Detail & Related papers (2023-03-30T08:15:18Z) - Multi-Prior Learning via Neural Architecture Search for Blind Face
Restoration [61.27907052910136]
Blind Face Restoration (BFR) aims to recover high-quality face images from low-quality ones.
Current methods still suffer from two major difficulties: 1) how to derive a powerful network architecture without extensive hand tuning; 2) how to capture complementary information from multiple facial priors in one network to improve restoration performance.
We propose a Face Restoration Searching Network (FRSNet) to adaptively search the suitable feature extraction architecture within our specified search space.
arXiv Detail & Related papers (2022-06-28T12:29:53Z) - Enhancing Quality of Pose-varied Face Restoration with Local Weak
Feature Sensing and GAN Prior [29.17397958948725]
We propose a well-designed blind face restoration network with generative facial prior.
Our model performs superior to the prior art for face restoration and face super-resolution tasks.
arXiv Detail & Related papers (2022-05-28T09:23:48Z) - FaceTuneGAN: Face Autoencoder for Convolutional Expression Transfer
Using Neural Generative Adversarial Networks [0.7043489166804575]
We present FaceTuneGAN, a new 3D face model representation decomposing and encoding separately facial identity and facial expression.
We propose a first adaptation of image-to-image translation networks, that have successfully been used in the 2D domain, to 3D face geometry.
arXiv Detail & Related papers (2021-12-01T14:42:03Z) - Subpixel Heatmap Regression for Facial Landmark Localization [65.41270740933656]
Heatmap regression approaches suffer from discretization-induced errors related to both the heatmap encoding and decoding process.
We propose a new approach for the heatmap encoding and decoding process by leveraging the underlying continuous distribution.
Our approach offers noticeable gains across multiple datasets setting a new state-of-the-art result in facial landmark localization.
arXiv Detail & Related papers (2021-11-03T17:21:28Z) - Inverting Generative Adversarial Renderer for Face Reconstruction [58.45125455811038]
In this work, we introduce a novel Generative Adversa Renderer (GAR)
GAR learns to model the complicated real-world image, instead of relying on the graphics rules, it is capable of producing realistic images.
Our method achieves state-of-the-art performances on multiple face reconstruction.
arXiv Detail & Related papers (2021-05-06T04:16:06Z) - High Resolution Face Editing with Masked GAN Latent Code Optimization [0.0]
Face editing is a popular research topic in the computer vision community.
Recent proposed methods are based on either training a conditional encoder-decoder Generative Adversarial Network (GAN) in an end-to-end fashion or on defining an operation in the latent space of a pre-trained vanilla GAN generator model.
We propose a GAN embedding optimization procedure with spatial and semantic constraints.
arXiv Detail & Related papers (2021-03-20T08:39:41Z) - The Power of Triply Complementary Priors for Image Compressive Sensing [89.14144796591685]
We propose a joint low-rank deep (LRD) image model, which contains a pair of complementaryly trip priors.
We then propose a novel hybrid plug-and-play framework based on the LRD model for image CS.
To make the optimization tractable, a simple yet effective algorithm is proposed to solve the proposed H-based image CS problem.
arXiv Detail & Related papers (2020-05-16T08:17:44Z) - Exploiting Deep Generative Prior for Versatile Image Restoration and
Manipulation [181.08127307338654]
This work presents an effective way to exploit the image prior captured by a generative adversarial network (GAN) trained on large-scale natural images.
The deep generative prior (DGP) provides compelling results to restore missing semantics, e.g., color, patch, resolution, of various degraded images.
arXiv Detail & Related papers (2020-03-30T17:45:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.