GANSpace: Discovering Interpretable GAN Controls
- URL: http://arxiv.org/abs/2004.02546v3
- Date: Mon, 14 Dec 2020 10:13:42 GMT
- Title: GANSpace: Discovering Interpretable GAN Controls
- Authors: Erik H\"ark\"onen, Aaron Hertzmann, Jaakko Lehtinen, Sylvain Paris
- Abstract summary: This paper describes a technique to analyze Generative Adversarial Networks (GANs) and create interpretable controls for image synthesis.
We identify important latent directions based on Principal Components Analysis (PCA) applied either in latent space or feature space.
We show that a large number of interpretable controls can be defined by layer-wise perturbation along the principal directions.
- Score: 24.428247009562895
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper describes a simple technique to analyze Generative Adversarial
Networks (GANs) and create interpretable controls for image synthesis, such as
change of viewpoint, aging, lighting, and time of day. We identify important
latent directions based on Principal Components Analysis (PCA) applied either
in latent space or feature space. Then, we show that a large number of
interpretable controls can be defined by layer-wise perturbation along the
principal directions. Moreover, we show that BigGAN can be controlled with
layer-wise inputs in a StyleGAN-like manner. We show results on different GANs
trained on various datasets, and demonstrate good qualitative matches to edit
directions found through earlier supervised approaches.
Related papers
- Discovering Interpretable Directions in the Semantic Latent Space of Diffusion Models [21.173910627285338]
Denoising Diffusion Models (DDMs) have emerged as a strong competitor to Generative Adversarial Networks (GANs)
In this paper, we explore the properties of h-space and propose several novel methods for finding meaningful semantic directions within it.
Our approaches are applicable without requiring architectural modifications, text-based guidance, CLIP-based optimization, or model fine-tuning.
arXiv Detail & Related papers (2023-03-20T12:59:32Z) - Discovering Class-Specific GAN Controls for Semantic Image Synthesis [73.91655061467988]
We propose a novel method for finding spatially disentangled class-specific directions in the latent space of pretrained SIS models.
We show that the latent directions found by our method can effectively control the local appearance of semantic classes.
arXiv Detail & Related papers (2022-12-02T21:39:26Z) - Exploring Gradient-based Multi-directional Controls in GANs [19.950198707910587]
We propose a novel approach that discovers nonlinear controls, which enables multi-directional manipulation as well as effective disentanglement.
Our approach is able to gain fine-grained controls over a diverse set of bi-directional and multi-directional attributes, and we showcase its ability to achieve disentanglement significantly better than state-of-the-art methods.
arXiv Detail & Related papers (2022-09-01T19:10:26Z) - PandA: Unsupervised Learning of Parts and Appearances in the Feature
Maps of GANs [34.145110544546114]
We present an architecture-agnostic approach that jointly discovers factors representing spatial parts and their appearances in an entirely unsupervised fashion.
Our method is far more efficient in terms of training time and, most importantly, provides much more accurate localized control.
arXiv Detail & Related papers (2022-05-31T18:28:39Z) - Fantastic Style Channels and Where to Find Them: A Submodular Framework
for Discovering Diverse Directions in GANs [0.0]
StyleGAN2 has enabled various image generation and manipulation tasks due to its rich and disentangled latent spaces.
We design a novel submodular framework that finds the most representative and diverse subset of directions in the latent space of StyleGAN2.
Our framework promotes diversity by using the notion of clusters and can be efficiently solved with a greedy optimization scheme.
arXiv Detail & Related papers (2022-03-16T10:35:41Z) - Unsupervised Discovery of Disentangled Manifolds in GANs [74.24771216154105]
Interpretable generation process is beneficial to various image editing applications.
We propose a framework to discover interpretable directions in the latent space given arbitrary pre-trained generative adversarial networks.
arXiv Detail & Related papers (2020-11-24T02:18:08Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z) - Closed-Form Factorization of Latent Semantics in GANs [65.42778970898534]
A rich set of interpretable dimensions has been shown to emerge in the latent space of the Generative Adversarial Networks (GANs) trained for synthesizing images.
In this work, we examine the internal representation learned by GANs to reveal the underlying variation factors in an unsupervised manner.
We propose a closed-form factorization algorithm for latent semantic discovery by directly decomposing the pre-trained weights.
arXiv Detail & Related papers (2020-07-13T18:05:36Z) - InterFaceGAN: Interpreting the Disentangled Face Representation Learned
by GANs [73.27299786083424]
We propose a framework called InterFaceGAN to interpret the disentangled face representation learned by state-of-the-art GAN models.
We first find that GANs learn various semantics in some linear subspaces of the latent space.
We then conduct a detailed study on the correlation between different semantics and manage to better disentangle them via subspace projection.
arXiv Detail & Related papers (2020-05-18T18:01:22Z) - Gradient-Induced Co-Saliency Detection [81.54194063218216]
Co-saliency detection (Co-SOD) aims to segment the common salient foreground in a group of relevant images.
In this paper, inspired by human behavior, we propose a gradient-induced co-saliency detection method.
arXiv Detail & Related papers (2020-04-28T08:40:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.