A naive method to discover directions in the StyleGAN2 latent space
- URL: http://arxiv.org/abs/2203.10373v1
- Date: Sat, 19 Mar 2022 18:43:16 GMT
- Title: A naive method to discover directions in the StyleGAN2 latent space
- Authors: Andrea Giardina, Soumya Subhra Paria, Adhikari Kaustubh
- Abstract summary: We show how the inversion process can be easily exploited to interpret the latent space and control the output of StyleGAN2, a GAN architecture capable of generating photo-realistic faces.
We show the results obtained by applying the proposed method to a set of photos extracted from the CelebA-HQ database.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Several research groups have shown that Generative Adversarial Networks
(GANs) can generate photo-realistic images in recent years. Using the GANs, a
map is created between a latent code and a photo-realistic image. This process
can also be reversed: given a photo as input, it is possible to obtain the
corresponding latent code. In this paper, we will show how the inversion
process can be easily exploited to interpret the latent space and control the
output of StyleGAN2, a GAN architecture capable of generating photo-realistic
faces. From a biological perspective, facial features such as nose size depend
on important genetic factors, and we explore the latent spaces that correspond
to such biological features, including masculinity and eye colour. We show the
results obtained by applying the proposed method to a set of photos extracted
from the CelebA-HQ database. We quantify some of these measures by utilizing
two landmarking protocols, and evaluate their robustness through statistical
analysis. Finally we correlate these measures with the input parameters used to
perturb the latent spaces along those interpretable directions. Our results
contribute towards building the groundwork of using such GAN architecture in
forensics to generate photo-realistic faces that satisfy certain biological
attributes.
Related papers
- Fluid Grey 2: How Well Does Generative Adversarial Network Learn Deeper Topology Structure in Architecture That Matches Images? [0.0]
This paper proves that pix2pix can automatically learn spatial topological relationships and apply them to architectural design.<n>It fills the gap in detecting the performance of Image-based Generation GAN from a topological perspective.<n>The two detection modules can be widely used for customizing image datasets with the same topological structure.
arXiv Detail & Related papers (2025-11-20T00:27:29Z) - Contrasting Deepfakes Diffusion via Contrastive Learning and Global-Local Similarities [88.398085358514]
Contrastive Deepfake Embeddings (CoDE) is a novel embedding space specifically designed for deepfake detection.
CoDE is trained via contrastive learning by additionally enforcing global-local similarities.
arXiv Detail & Related papers (2024-07-29T18:00:10Z) - GH-Feat: Learning Versatile Generative Hierarchical Features from GANs [61.208757845344074]
We show that a generative feature learned from image synthesis exhibits great potentials in solving a wide range of computer vision tasks.
We first train an encoder by considering the pretrained StyleGAN generator as a learned loss function.
The visual features produced by our encoder, termed as Generative Hierarchical Features (GH-Feat), highly align with the layer-wise GAN representations.
arXiv Detail & Related papers (2023-01-12T21:59:46Z) - Detecting High-Quality GAN-Generated Face Images using Neural Networks [23.388645531702597]
We propose a new strategy to differentiate GAN-generated images from authentic images by leveraging spectral band discrepancies.
In particular, we enable the digital preservation of face images using the Cross-band co-occurrence matrix and spatial co-occurrence matrix.
We show that the performance boost is particularly significant and achieves more than 92% in different post-processing environments.
arXiv Detail & Related papers (2022-03-03T13:53:27Z) - Low-Rank Subspaces in GANs [101.48350547067628]
This work introduces low-rank subspaces that enable more precise control of GAN generation.
LowRankGAN is able to find the low-dimensional representation of attribute manifold.
Experiments on state-of-the-art GAN models (including StyleGAN2 and BigGAN) trained on various datasets demonstrate the effectiveness of our LowRankGAN.
arXiv Detail & Related papers (2021-06-08T16:16:32Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z) - The Geometry of Deep Generative Image Models and its Applications [0.0]
Generative adversarial networks (GANs) have emerged as a powerful unsupervised method to model the statistical patterns of real-world data sets.
These networks are trained to map random inputs in their latent space to new samples representative of the learned data.
The structure of the latent space is hard to intuit due to its high dimensionality and the non-linearity of the generator.
arXiv Detail & Related papers (2021-01-15T07:57:33Z) - CNN Detection of GAN-Generated Face Images based on Cross-Band
Co-occurrences Analysis [34.41021278275805]
Last-generation GAN models allow to generate synthetic images which are visually indistinguishable from natural ones.
We propose a method for distinguishing GAN-generated from natural images by exploiting inconsistencies among spectral bands.
arXiv Detail & Related papers (2020-07-25T10:55:04Z) - InterFaceGAN: Interpreting the Disentangled Face Representation Learned
by GANs [73.27299786083424]
We propose a framework called InterFaceGAN to interpret the disentangled face representation learned by state-of-the-art GAN models.
We first find that GANs learn various semantics in some linear subspaces of the latent space.
We then conduct a detailed study on the correlation between different semantics and manage to better disentangle them via subspace projection.
arXiv Detail & Related papers (2020-05-18T18:01:22Z) - Deep Snow: Synthesizing Remote Sensing Imagery with Generative
Adversarial Nets [0.5249805590164901]
generative adversarial networks (GANs) can be used to generate realistic pervasive changes in remote sensing imagery.
We investigate some transformation quality metrics based on deep embedding of the generated and real images.
arXiv Detail & Related papers (2020-05-18T17:05:00Z) - Pathological Retinal Region Segmentation From OCT Images Using Geometric
Relation Based Augmentation [84.7571086566595]
We propose improvements over previous GAN-based medical image synthesis methods by jointly encoding the intrinsic relationship of geometry and shape.
The proposed method outperforms state-of-the-art segmentation methods on the public RETOUCH dataset having images captured from different acquisition procedures.
arXiv Detail & Related papers (2020-03-31T11:50:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.