Disentangled Lifespan Face Synthesis
- URL: http://arxiv.org/abs/2108.02874v1
- Date: Thu, 5 Aug 2021 22:33:14 GMT
- Title: Disentangled Lifespan Face Synthesis
- Authors: Sen He, Wentong Liao, Michael Ying Yang, Yi-Zhe Song, Bodo Rosenhahn,
Tao Xiang
- Abstract summary: A lifespan face synthesis (LFS) model aims to generate a set of photo-realistic face images of a person's whole life, given only one snapshot as reference.
The generated face image given a target age code is expected to be age-sensitive reflected by bio-plausible transformations of shape and texture.
This is achieved by extracting shape, texture and identity features separately from an encoder.
- Score: 100.29058545878341
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A lifespan face synthesis (LFS) model aims to generate a set of
photo-realistic face images of a person's whole life, given only one snapshot
as reference. The generated face image given a target age code is expected to
be age-sensitive reflected by bio-plausible transformations of shape and
texture, while being identity preserving. This is extremely challenging because
the shape and texture characteristics of a face undergo separate and highly
nonlinear transformations w.r.t. age. Most recent LFS models are based on
generative adversarial networks (GANs) whereby age code conditional
transformations are applied to a latent face representation. They benefit
greatly from the recent advancements of GANs. However, without explicitly
disentangling their latent representations into the texture, shape and identity
factors, they are fundamentally limited in modeling the nonlinear age-related
transformation on texture and shape whilst preserving identity. In this work, a
novel LFS model is proposed to disentangle the key face characteristics
including shape, texture and identity so that the unique shape and texture age
transformations can be modeled effectively. This is achieved by extracting
shape, texture and identity features separately from an encoder. Critically,
two transformation modules, one conditional convolution based and the other
channel attention based, are designed for modeling the nonlinear shape and
texture feature transformations respectively. This is to accommodate their
rather distinct aging processes and ensure that our synthesized images are both
age-sensitive and identity preserving. Extensive experiments show that our LFS
model is clearly superior to the state-of-the-art alternatives. Codes and demo
are available on our project website:
\url{https://senhe.github.io/projects/iccv_2021_lifespan_face}.
Related papers
- G2Face: High-Fidelity Reversible Face Anonymization via Generative and Geometric Priors [71.69161292330504]
Reversible face anonymization seeks to replace sensitive identity information in facial images with synthesized alternatives.
This paper introduces Gtextsuperscript2Face, which leverages both generative and geometric priors to enhance identity manipulation.
Our method outperforms existing state-of-the-art techniques in face anonymization and recovery, while preserving high data utility.
arXiv Detail & Related papers (2024-08-18T12:36:47Z) - When StyleGAN Meets Stable Diffusion: a $\mathscr{W}_+$ Adapter for
Personalized Image Generation [60.305112612629465]
Text-to-image diffusion models have excelled in producing diverse, high-quality, and photo-realistic images.
We present a novel use of the extended StyleGAN embedding space $mathcalW_+$ to achieve enhanced identity preservation and disentanglement for diffusion models.
Our method adeptly generates personalized text-to-image outputs that are not only compatible with prompt descriptions but also amenable to common StyleGAN editing directions.
arXiv Detail & Related papers (2023-11-29T09:05:14Z) - Learning to Aggregate and Personalize 3D Face from In-the-Wild Photo
Collection [65.92058628082322]
Non-parametric face modeling aims to reconstruct 3D face only from images without shape assumptions.
This paper presents a novel Learning to Aggregate and Personalize framework for unsupervised robust 3D face modeling.
arXiv Detail & Related papers (2021-06-15T03:10:17Z) - AgeFlow: Conditional Age Progression and Regression with Normalizing
Flows [19.45760984401544]
Age progression and regression aim to synthesize photorealistic appearance of a given face image with aging and rejuvenation effects, respectively.
Existing generative adversarial networks (GANs) based methods suffer from the following three major issues.
This paper proposes a novel framework, termed AgeFlow, to integrate the advantages of both flow-based models and GANs.
arXiv Detail & Related papers (2021-05-15T15:02:07Z) - When Age-Invariant Face Recognition Meets Face Age Synthesis: A
Multi-Task Learning Framework [20.579282497730944]
MTLFace can learn age-invariant identity-related representation while achieving pleasing face synthesis.
In contrast to the conventional one-hot encoding that achieves group-level FAS, we propose a novel identity conditional module to achieve identity-level FAS.
Extensive experiments on five benchmark cross-age datasets demonstrate the superior performance of our proposed MTLFace.
arXiv Detail & Related papers (2021-03-02T07:03:27Z) - Continuous Face Aging Generative Adversarial Networks [11.75204350455584]
Face aging is the task aiming to translate the faces in input images to designated ages.
Previous methods have limited themselves only able to produce discrete age groups, each of which consists of ten years.
We propose the continuous face aging generative adversarial networks (CFA-GAN)
arXiv Detail & Related papers (2021-02-26T06:22:25Z) - Progressive Semantic-Aware Style Transformation for Blind Face
Restoration [26.66332852514812]
We propose a new progressive semantic-aware style transformation framework, named PSFR-GAN, for face restoration.
The proposed PSFR-GAN makes full use of the semantic (parsing maps) and pixel (LQ images) space information from different scales of input pairs.
Experiment results show that our model trained with synthetic data can not only produce more realistic high-resolution results for synthetic LQ inputs but also better to generalize natural LQ face images.
arXiv Detail & Related papers (2020-09-18T09:27:33Z) - DotFAN: A Domain-transferred Face Augmentation Network for Pose and
Illumination Invariant Face Recognition [94.96686189033869]
We propose a 3D model-assisted domain-transferred face augmentation network (DotFAN)
DotFAN can generate a series of variants of an input face based on the knowledge distilled from existing rich face datasets collected from other domains.
Experiments show that DotFAN is beneficial for augmenting small face datasets to improve their within-class diversity.
arXiv Detail & Related papers (2020-02-23T08:16:34Z) - Fine-grained Image-to-Image Transformation towards Visual Recognition [102.51124181873101]
We aim at transforming an image with a fine-grained category to synthesize new images that preserve the identity of the input image.
We adopt a model based on generative adversarial networks to disentangle the identity related and unrelated factors of an image.
Experiments on the CompCars and Multi-PIE datasets demonstrate that our model preserves the identity of the generated images much better than the state-of-the-art image-to-image transformation models.
arXiv Detail & Related papers (2020-01-12T05:26:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.