Information-theoretic stochastic contrastive conditional GAN:
InfoSCC-GAN
- URL: http://arxiv.org/abs/2112.09653v1
- Date: Fri, 17 Dec 2021 17:56:30 GMT
- Title: Information-theoretic stochastic contrastive conditional GAN:
InfoSCC-GAN
- Authors: Vitaliy Kinakh, Mariia Drozdova, Guillaume Qu\'etant, Tobias Golling,
Slava Voloshynovskiy
- Abstract summary: We present a contrastive conditional generative adversarial network (Info SCC-GAN) with an explorable latent space.
Info SCC-GAN is derived based on an information-theoretic formulation of mutual information between input data and latent space representation.
Experiments show that Info SCC-GAN outperforms the "vanilla" EigenGAN in the image generation on AFHQ and CelebA datasets.
- Score: 6.201770337181472
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conditional generation is a subclass of generative problems where the output
of the generation is conditioned by the attribute information. In this paper,
we present a stochastic contrastive conditional generative adversarial network
(InfoSCC-GAN) with an explorable latent space. The InfoSCC-GAN architecture is
based on an unsupervised contrastive encoder built on the InfoNCE paradigm, an
attribute classifier and an EigenGAN generator. We propose a novel training
method, based on generator regularization using external or internal attributes
every $n$-th iteration, using a pre-trained contrastive encoder and a
pre-trained classifier. The proposed InfoSCC-GAN is derived based on an
information-theoretic formulation of mutual information maximization between
input data and latent space representation as well as latent space and
generated data. Thus, we demonstrate a link between the training objective
functions and the above information-theoretic formulation. The experimental
results show that InfoSCC-GAN outperforms the "vanilla" EigenGAN in the image
generation on AFHQ and CelebA datasets. In addition, we investigate the impact
of discriminator architectures and loss functions by performing ablation
studies. Finally, we demonstrate that thanks to the EigenGAN generator, the
proposed framework enjoys a stochastic generation in contrast to vanilla
deterministic GANs yet with the independent training of encoder, classifier,
and generator in contrast to existing frameworks. Code, experimental results,
and demos are available online at https://github.com/vkinakh/InfoSCC-GAN.
Related papers
- Disentanglement via Latent Quantization [60.37109712033694]
In this work, we construct an inductive bias towards encoding to and decoding from an organized latent space.
We demonstrate the broad applicability of this approach by adding it to both basic data-re (vanilla autoencoder) and latent-reconstructing (InfoGAN) generative models.
arXiv Detail & Related papers (2023-05-28T06:30:29Z) - Sequential training of GANs against GAN-classifiers reveals correlated
"knowledge gaps" present among independently trained GAN instances [1.104121146441257]
We iteratively train GAN-classifiers and train GANs that "fool" the classifiers.
We examine the effect on GAN training dynamics, output quality, and GAN-classifier generalization.
arXiv Detail & Related papers (2023-03-27T18:18:15Z) - Collapse by Conditioning: Training Class-conditional GANs with Limited
Data [109.30895503994687]
We propose a training strategy for conditional GANs (cGANs) that effectively prevents the observed mode-collapse by leveraging unconditional learning.
Our training strategy starts with an unconditional GAN and gradually injects conditional information into the generator and the objective function.
The proposed method for training cGANs with limited data results not only in stable training but also in generating high-quality images.
arXiv Detail & Related papers (2022-01-17T18:59:23Z) - Improving Model Compatibility of Generative Adversarial Networks by
Boundary Calibration [24.28407308818025]
Boundary-Calibration GANs (BCGANs) are proposed to improve GAN's model compatibility.
BCGANs generate realistic images like original GANs but also achieves superior model compatibility than the original GANs.
arXiv Detail & Related papers (2021-11-03T16:08:09Z) - Improving the quality of generative models through Smirnov
transformation [1.3492000366723798]
We propose a novel activation function to be used as output of the generator agent.
It is based on the Smirnov probabilistic transformation and it is specifically designed to improve the quality of the generated data.
arXiv Detail & Related papers (2021-10-29T17:01:06Z) - cGANs with Auxiliary Discriminative Classifier [43.78253518292111]
Conditional generative models aim to learn the underlying joint distribution of data and labels.
auxiliary classifier generative adversarial networks (AC-GAN) have been widely used, but suffer from the issue of low intra-class diversity on generated samples.
We propose novel cGANs with auxiliary discriminative classifier (ADC-GAN) to address the issue of AC-GAN.
arXiv Detail & Related papers (2021-07-21T13:06:32Z) - Conditional Hybrid GAN for Sequence Generation [56.67961004064029]
We propose a novel conditional hybrid GAN (C-Hybrid-GAN) to solve this issue.
We exploit the Gumbel-Softmax technique to approximate the distribution of discrete-valued sequences.
We demonstrate that the proposed C-Hybrid-GAN outperforms the existing methods in context-conditioned discrete-valued sequence generation.
arXiv Detail & Related papers (2020-09-18T03:52:55Z) - Improving Generative Adversarial Networks with Local Coordinate Coding [150.24880482480455]
Generative adversarial networks (GANs) have shown remarkable success in generating realistic data from some predefined prior distribution.
In practice, semantic information might be represented by some latent distribution learned from data.
We propose an LCCGAN model with local coordinate coding (LCC) to improve the performance of generating data.
arXiv Detail & Related papers (2020-07-28T09:17:50Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z) - Inducing Optimal Attribute Representations for Conditional GANs [61.24506213440997]
Conditional GANs are widely used in translating an image from one category to another.
Existing conditional GANs commonly encode target domain label information as hard-coded categorical vectors in the form of 0s and 1s.
We propose a novel end-to-end learning framework with Graph Convolutional Networks to learn the attribute representations to condition on the generator.
arXiv Detail & Related papers (2020-03-13T20:24:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.