Heredity-aware Child Face Image Generation with Latent Space
Disentanglement
- URL: http://arxiv.org/abs/2108.11080v1
- Date: Wed, 25 Aug 2021 06:59:43 GMT
- Title: Heredity-aware Child Face Image Generation with Latent Space
Disentanglement
- Authors: Xiao Cui, Wengang Zhou, Yang Hu, Weilun Wang and Houqiang Li
- Abstract summary: We propose a novel approach, called ChildGAN, to generate a child's image according to the images of parents with heredity prior.
The main idea is to disentangle the latent space of a pre-trained generation model and precisely control the face attributes of child images with clear semantics.
- Score: 96.92684978356425
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative adversarial networks have been widely used in image synthesis in
recent years and the quality of the generated image has been greatly improved.
However, the flexibility to control and decouple facial attributes (e.g., eyes,
nose, mouth) is still limited. In this paper, we propose a novel approach,
called ChildGAN, to generate a child's image according to the images of parents
with heredity prior. The main idea is to disentangle the latent space of a
pre-trained generation model and precisely control the face attributes of child
images with clear semantics. We use distances between face landmarks as pseudo
labels to figure out the most influential semantic vectors of the corresponding
face attributes by calculating the gradient of latent vectors to pseudo labels.
Furthermore, we disentangle the semantic vectors by weighting irrelevant
features and orthogonalizing them with Schmidt Orthogonalization. Finally, we
fuse the latent vector of the parents by leveraging the disentangled semantic
vectors under the guidance of biological genetic laws. Extensive experiments
demonstrate that our approach outperforms the existing methods with encouraging
results.
Related papers
- Semantic Contextualization of Face Forgery: A New Definition, Dataset, and Detection Method [77.65459419417533]
We put face forgery in a semantic context and define that computational methods that alter semantic face attributes are sources of face forgery.
We construct a large face forgery image dataset, where each image is associated with a set of labels organized in a hierarchical graph.
We propose a semantics-oriented face forgery detection method that captures label relations and prioritizes the primary task.
arXiv Detail & Related papers (2024-05-14T10:24:19Z) - Attribute Controllable Beautiful Caucasian Face Generation by Aesthetics
Driven Reinforcement Learning [21.329906392100884]
We build the techniques of reinforcement learning into the generator of EigenGAN.
The agent tries to figure out how to alter the semantic attributes of the generated human faces towards more preferable ones.
We present a new variant incorporating the ingredients emerging in the reinforcement learning communities in recent years.
arXiv Detail & Related papers (2022-08-09T03:04:10Z) - Semantic Unfolding of StyleGAN Latent Space [0.7646713951724012]
Generative adversarial networks (GANs) have proven to be surprisingly efficient for image editing by inverting and manipulating the latent code corresponding to an input real image.
This editing property emerges from the disentangled nature of the latent space.
In this paper, we identify that the facial attribute disentanglement is not optimal, thus facial editing relying on linear attribute separation is flawed.
arXiv Detail & Related papers (2022-06-29T20:22:10Z) - Spatially Multi-conditional Image Generation [80.04130168156792]
We propose a novel neural architecture to address the problem of multi-conditional image generation.
The proposed method uses a transformer-like architecture operating pixel-wise, which receives the available labels as input tokens.
Our experiments on three benchmark datasets demonstrate the clear superiority of our method over the state-of-the-art and the compared baselines.
arXiv Detail & Related papers (2022-03-25T17:57:13Z) - Semantic Segmentation with Generative Models: Semi-Supervised Learning
and Strong Out-of-Domain Generalization [112.68171734288237]
We propose a novel framework for discriminative pixel-level tasks using a generative model of both images and labels.
We learn a generative adversarial network that captures the joint image-label distribution and is trained efficiently using a large set of unlabeled images.
We demonstrate strong in-domain performance compared to several baselines, and are the first to showcase extreme out-of-domain generalization.
arXiv Detail & Related papers (2021-04-12T21:41:25Z) - Unsupervised Discovery of Disentangled Manifolds in GANs [74.24771216154105]
Interpretable generation process is beneficial to various image editing applications.
We propose a framework to discover interpretable directions in the latent space given arbitrary pre-trained generative adversarial networks.
arXiv Detail & Related papers (2020-11-24T02:18:08Z) - InterFaceGAN: Interpreting the Disentangled Face Representation Learned
by GANs [73.27299786083424]
We propose a framework called InterFaceGAN to interpret the disentangled face representation learned by state-of-the-art GAN models.
We first find that GANs learn various semantics in some linear subspaces of the latent space.
We then conduct a detailed study on the correlation between different semantics and manage to better disentangle them via subspace projection.
arXiv Detail & Related papers (2020-05-18T18:01:22Z) - Disentangled and Controllable Face Image Generation via 3D
Imitative-Contrastive Learning [43.53235319568048]
We propose DiscoFaceGAN, an approach for face image generation of virtual people with disentangled, precisely-controllable latent representations.
We embed 3D priors into adversarial learning and train the network to imitate the image formation of an analytic 3D face deformation and rendering process.
arXiv Detail & Related papers (2020-04-24T11:21:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.