DO-GAN: A Double Oracle Framework for Generative Adversarial Networks
- URL: http://arxiv.org/abs/2102.08577v1
- Date: Wed, 17 Feb 2021 05:11:18 GMT
- Title: DO-GAN: A Double Oracle Framework for Generative Adversarial Networks
- Authors: Aye Phyu Phyu Aung, Xinrun Wang, Runsheng Yu, Bo An, Senthilnath
Jayavelu, Xiaoli Li
- Abstract summary: We propose a new approach to train Generative Adversarial Networks (GANs)
We deploy a double-oracle framework using the generator and discriminator oracles.
We apply our framework to established GAN architectures such as vanilla GAN, Deep Convolutional GAN, Spectral Normalization GAN and Stacked GAN.
- Score: 28.904057977044374
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a new approach to train Generative Adversarial
Networks (GANs) where we deploy a double-oracle framework using the generator
and discriminator oracles. GAN is essentially a two-player zero-sum game
between the generator and the discriminator. Training GANs is challenging as a
pure Nash equilibrium may not exist and even finding the mixed Nash equilibrium
is difficult as GANs have a large-scale strategy space. In DO-GAN, we extend
the double oracle framework to GANs. We first generalize the players'
strategies as the trained models of generator and discriminator from the best
response oracles. We then compute the meta-strategies using a linear program.
For scalability of the framework where multiple generators and discriminator
best responses are stored in the memory, we propose two solutions: 1) pruning
the weakly-dominated players' strategies to keep the oracles from becoming
intractable; 2) applying continual learning to retain the previous knowledge of
the networks. We apply our framework to established GAN architectures such as
vanilla GAN, Deep Convolutional GAN, Spectral Normalization GAN and Stacked
GAN. Finally, we conduct experiments on MNIST, CIFAR-10 and CelebA datasets and
show that DO-GAN variants have significant improvements in both subjective
qualitative evaluation and quantitative metrics, compared with their respective
GAN architectures.
Related papers
- FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - Double Oracle Neural Architecture Search for Game Theoretic Deep Learning Models [28.238075755838487]
We propose a new approach to train deep learning models using game theory concepts.
We deploy a double-versarial framework using best response oracles.
We show that all our variants have significant improvements in both subjective qualitative evaluation and quantitative metrics.
arXiv Detail & Related papers (2024-10-07T05:42:01Z) - LD-GAN: Low-Dimensional Generative Adversarial Network for Spectral
Image Generation with Variance Regularization [72.4394510913927]
Deep learning methods are state-of-the-art for spectral image (SI) computational tasks.
GANs enable diverse augmentation by learning and sampling from the data distribution.
GAN-based SI generation is challenging since the high-dimensionality nature of this kind of data hinders the convergence of the GAN training yielding to suboptimal generation.
We propose a statistical regularization to control the low-dimensional representation variance for the autoencoder training and to achieve high diversity of samples generated with the GAN.
arXiv Detail & Related papers (2023-04-29T00:25:02Z) - Self-Ensembling GAN for Cross-Domain Semantic Segmentation [107.27377745720243]
This paper proposes a self-ensembling generative adversarial network (SE-GAN) exploiting cross-domain data for semantic segmentation.
In SE-GAN, a teacher network and a student network constitute a self-ensembling model for generating semantic segmentation maps, which together with a discriminator, forms a GAN.
Despite its simplicity, we find SE-GAN can significantly boost the performance of adversarial training and enhance the stability of the model.
arXiv Detail & Related papers (2021-12-15T09:50:25Z) - Fuzzy Generative Adversarial Networks [0.0]
Generative Adversarial Networks (GANs) are well-known tools for data generation and semi-supervised classification.
This paper introduces techniques that show improvement in the GANs' regression capability through mean absolute error (MAE) and mean squared error (MSE)
We show that adding a fuzzy logic layer can enhance GAN's ability to perform regression; the most desirable injection location is problem-specific.
arXiv Detail & Related papers (2021-10-27T17:05:06Z) - Inferential Wasserstein Generative Adversarial Networks [9.859829604054127]
We introduce a novel inferential Wasserstein GAN (iWGAN) model, which is a principled framework to fuse auto-encoders and WGANs.
The iWGAN greatly mitigates the symptom of mode collapse, speeds up the convergence, and is able to provide a measurement of quality check for each individual sample.
arXiv Detail & Related papers (2021-09-13T00:43:21Z) - Class Balancing GAN with a Classifier in the Loop [58.29090045399214]
We introduce a novel theoretically motivated Class Balancing regularizer for training GANs.
Our regularizer makes use of the knowledge from a pre-trained classifier to ensure balanced learning of all the classes in the dataset.
We demonstrate the utility of our regularizer in learning representations for long-tailed distributions via achieving better performance than existing approaches over multiple datasets.
arXiv Detail & Related papers (2021-06-17T11:41:30Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z) - Robust Generative Adversarial Network [37.015223009069175]
We aim to improve the generalization capability of GANs by promoting the local robustness within the small neighborhood of the training samples.
We design a robust optimization framework where the generator and discriminator compete with each other in a textitworst-case setting within a small Wasserstein ball.
We have proved that our robust method can obtain a tighter generalization upper bound than traditional GANs under mild assumptions.
arXiv Detail & Related papers (2020-04-28T07:37:01Z) - Discriminator Contrastive Divergence: Semi-Amortized Generative Modeling
by Exploring Energy of the Discriminator [85.68825725223873]
Generative Adversarial Networks (GANs) have shown great promise in modeling high dimensional data.
We introduce the Discriminator Contrastive Divergence, which is well motivated by the property of WGAN's discriminator.
We demonstrate the benefits of significant improved generation on both synthetic data and several real-world image generation benchmarks.
arXiv Detail & Related papers (2020-04-05T01:50:16Z) - xAI-GAN: Enhancing Generative Adversarial Networks via Explainable AI
Systems [16.360144499713524]
Generative Adversarial Networks (GANs) are a revolutionary class of Deep Neural Networks (DNNs) that have been successfully used to generate realistic images, music, text, and other data.
We propose a new class of GAN that leverages recent advances in explainable AI (xAI) systems to provide a "richer" form of corrective feedback from discriminators to generators.
We observe xAI-GANs provide an improvement of up to 23.18% in the quality of generated images on both MNIST and FMNIST datasets over standard GANs.
arXiv Detail & Related papers (2020-02-24T18:38:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.