AE-OT-GAN: Training GANs from data specific latent distribution
- URL: http://arxiv.org/abs/2001.03698v2
- Date: Mon, 27 Jan 2020 15:25:28 GMT
- Title: AE-OT-GAN: Training GANs from data specific latent distribution
- Authors: Dongsheng An, Yang Guo, Min Zhang, Xin Qi, Na Lei, Shing-Tung Yau, and
Xianfeng Gu
- Abstract summary: generative adversarial networks (GANs) areprominent models to generate realistic and crisp images.
GANs often encounter the mode collapse problems and arehard to train, which comes from approximating the intrinsicdiscontinuous distribution transform map with continuousDNNs.
The recently proposed AE-OT model addresses thisproblem by explicitly computing the discontinuous distribu-tion transform map.
In this paper, wepropose the AE-OT-GAN model to utilize the advantages ofthe both models: generate high quality images and at the same time overcome the mode collapse/mixture problems.
- Score: 21.48007565143911
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Though generative adversarial networks (GANs) areprominent models to generate
realistic and crisp images,they often encounter the mode collapse problems and
arehard to train, which comes from approximating the intrinsicdiscontinuous
distribution transform map with continuousDNNs. The recently proposed AE-OT
model addresses thisproblem by explicitly computing the discontinuous
distribu-tion transform map through solving a semi-discrete optimaltransport
(OT) map in the latent space of the autoencoder.However the generated images
are blurry. In this paper, wepropose the AE-OT-GAN model to utilize the
advantages ofthe both models: generate high quality images and at thesame time
overcome the mode collapse/mixture problems.Specifically, we first faithfully
embed the low dimensionalimage manifold into the latent space by training an
autoen-coder (AE). Then we compute the optimal transport (OT)map that pushes
forward the uniform distribution to the la-tent distribution supported on the
latent manifold. Finally,our GAN model is trained to generate high quality
imagesfrom the latent distribution, the distribution transform mapfrom which to
the empirical data distribution will be con-tinuous. The paired data between
the latent code and thereal images gives us further constriction about the
generator.Experiments on simple MNIST dataset and complex datasetslike Cifar-10
and CelebA show the efficacy and efficiency ofour proposed method.
Related papers
- Learning Gaussian Representation for Eye Fixation Prediction [54.88001757991433]
Existing eye fixation prediction methods perform the mapping from input images to the corresponding dense fixation maps generated from raw fixation points.
We introduce Gaussian Representation for eye fixation modeling.
We design our framework upon some lightweight backbones to achieve real-time fixation prediction.
arXiv Detail & Related papers (2024-03-21T20:28:22Z) - Denoising Diffusion Bridge Models [54.87947768074036]
Diffusion models are powerful generative models that map noise to data using processes.
For many applications such as image editing, the model input comes from a distribution that is not random noise.
In our work, we propose Denoising Diffusion Bridge Models (DDBMs)
arXiv Detail & Related papers (2023-09-29T03:24:24Z) - Complexity Matters: Rethinking the Latent Space for Generative Modeling [65.64763873078114]
In generative modeling, numerous successful approaches leverage a low-dimensional latent space, e.g., Stable Diffusion.
In this study, we aim to shed light on this under-explored topic by rethinking the latent space from the perspective of model complexity.
arXiv Detail & Related papers (2023-07-17T07:12:29Z) - Hierarchical Integration Diffusion Model for Realistic Image Deblurring [71.76410266003917]
Diffusion models (DMs) have been introduced in image deblurring and exhibited promising performance.
We propose the Hierarchical Integration Diffusion Model (HI-Diff), for realistic image deblurring.
Experiments on synthetic and real-world blur datasets demonstrate that our HI-Diff outperforms state-of-the-art methods.
arXiv Detail & Related papers (2023-05-22T12:18:20Z) - DDFM: Denoising Diffusion Model for Multi-Modality Image Fusion [144.9653045465908]
We propose a novel fusion algorithm based on the denoising diffusion probabilistic model (DDPM)
Our approach yields promising fusion results in infrared-visible image fusion and medical image fusion.
arXiv Detail & Related papers (2023-03-13T04:06:42Z) - Latent Space is Feature Space: Regularization Term for GANs Training on
Limited Dataset [1.8634083978855898]
I proposed an additional structure and loss function for GANs called LFM, trained to maximize the feature diversity between the different dimensions of the latent space.
In experiments, this system has been built upon DCGAN and proved to have improvement on Frechet Inception Distance (FID) training from scratch on CelebA dataset.
arXiv Detail & Related papers (2022-10-28T16:34:48Z) - Image Generation with Multimodal Priors using Denoising Diffusion
Probabilistic Models [54.1843419649895]
A major challenge in using generative models to accomplish this task is the lack of paired data containing all modalities and corresponding outputs.
We propose a solution based on a denoising diffusion probabilistic synthesis models to generate images under multi-model priors.
arXiv Detail & Related papers (2022-06-10T12:23:05Z) - Learning Multiple Probabilistic Degradation Generators for Unsupervised
Real World Image Super Resolution [5.987801889633082]
Unsupervised real world super resolution aims at restoring high-resolution (HR) images given low-resolution (LR) inputs when paired data is unavailable.
One of the most common approaches is synthesizing noisy LR images using GANs and utilizing a synthetic dataset to train the model in a supervised manner.
We propose a probabilistic degradation generator to approximate the distribution of LR images given a HR image.
arXiv Detail & Related papers (2022-01-26T04:49:11Z) - Generation of data on discontinuous manifolds via continuous stochastic
non-invertible networks [6.201770337181472]
We show how to generate discontinuous distributions using continuous networks.
We derive a link between the cost functions and the information-theoretic formulation.
We apply our approach to synthetic 2D distributions to demonstrate both reconstruction and generation of discontinuous distributions.
arXiv Detail & Related papers (2021-12-17T17:39:59Z) - Generative Model without Prior Distribution Matching [26.91643368299913]
Variational Autoencoder (VAE) and its variations are classic generative models by learning a low-dimensional latent representation to satisfy some prior distribution.
We propose to let the prior match the embedding distribution rather than imposing the latent variables to fit the prior.
arXiv Detail & Related papers (2020-09-23T09:33:24Z) - Optimizing Generative Adversarial Networks for Image Super Resolution
via Latent Space Regularization [4.529132742139768]
Generative Adversarial Networks (GANs) try to learn the distribution of the real images in the manifold to generate samples that look real.
We probe for ways to alleviate these problems for supervised GANs in this paper.
arXiv Detail & Related papers (2020-01-22T16:27:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.