A Unified View of cGANs with and without Classifiers
- URL: http://arxiv.org/abs/2111.01035v1
- Date: Mon, 1 Nov 2021 15:36:33 GMT
- Title: A Unified View of cGANs with and without Classifiers
- Authors: Si-An Chen, Chun-Liang Li, Hsuan-Tien Lin
- Abstract summary: Conditional Generative Adversarial Networks (cGANs) are implicit generative models which allow to sample from class-conditional distributions.
Some representative cGANs avoid the shortcoming and reach state-of-the-art performance without having classifiers.
In this work, we demonstrate that classifiers can be properly leveraged to improve cGANs.
- Score: 24.28407308818025
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Conditional Generative Adversarial Networks (cGANs) are implicit generative
models which allow to sample from class-conditional distributions. Existing
cGANs are based on a wide range of different discriminator designs and training
objectives. One popular design in earlier works is to include a classifier
during training with the assumption that good classifiers can help eliminate
samples generated with wrong classes. Nevertheless, including classifiers in
cGANs often comes with a side effect of only generating easy-to-classify
samples. Recently, some representative cGANs avoid the shortcoming and reach
state-of-the-art performance without having classifiers. Somehow it remains
unanswered whether the classifiers can be resurrected to design better cGANs.
In this work, we demonstrate that classifiers can be properly leveraged to
improve cGANs. We start by using the decomposition of the joint probability
distribution to connect the goals of cGANs and classification as a unified
framework. The framework, along with a classic energy model to parameterize
distributions, justifies the use of classifiers for cGANs in a principled
manner. It explains several popular cGAN variants, such as ACGAN, ProjGAN, and
ContraGAN, as special cases with different levels of approximations, which
provides a unified view and brings new insights to understanding cGANs.
Experimental results demonstrate that the design inspired by the proposed
framework outperforms state-of-the-art cGANs on multiple benchmark datasets,
especially on the most challenging ImageNet. The code is available at
https://github.com/sian-chen/PyTorch-ECGAN.
Related papers
Err
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.