Combating Mode Collapse in GANs via Manifold Entropy Estimation
- URL: http://arxiv.org/abs/2208.12055v6
- Date: Sat, 8 Apr 2023 11:03:02 GMT
- Title: Combating Mode Collapse in GANs via Manifold Entropy Estimation
- Authors: Haozhe Liu, Bing Li, Haoqian Wu, Hanbang Liang, Yawen Huang, Yuexiang
Li, Bernard Ghanem, Yefeng Zheng
- Abstract summary: Generative Adversarial Networks (GANs) have shown compelling results in various tasks and applications.
We propose a novel training pipeline to address the mode collapse issue of GANs.
- Score: 70.06639443446545
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative Adversarial Networks (GANs) have shown compelling results in
various tasks and applications in recent years. However, mode collapse remains
a critical problem in GANs. In this paper, we propose a novel training pipeline
to address the mode collapse issue of GANs. Different from existing methods, we
propose to generalize the discriminator as feature embedding and maximize the
entropy of distributions in the embedding space learned by the discriminator.
Specifically, two regularization terms, i.e., Deep Local Linear Embedding
(DLLE) and Deep Isometric feature Mapping (DIsoMap), are designed to encourage
the discriminator to learn the structural information embedded in the data,
such that the embedding space learned by the discriminator can be well-formed.
Based on the well-learned embedding space supported by the discriminator, a
non-parametric entropy estimator is designed to efficiently maximize the
entropy of embedding vectors, playing as an approximation of maximizing the
entropy of the generated distribution. By improving the discriminator and
maximizing the distance of the most similar samples in the embedding space, our
pipeline effectively reduces the mode collapse without sacrificing the quality
of generated samples. Extensive experimental results show the effectiveness of
our method, which outperforms the GAN baseline, MaF-GAN on CelebA (9.13 vs.
12.43 in FID) and surpasses the recent state-of-the-art energy-based model on
the ANIME-FACE dataset (2.80 vs. 2.26 in Inception score). The code is
available at https://github.com/HaozheLiu-ST/MEE
Related papers
- Towards Robust Out-of-Distribution Generalization: Data Augmentation and Neural Architecture Search Approaches [4.577842191730992]
We study ways toward robust OoD generalization for deep learning.
We first propose a novel and effective approach to disentangle the spurious correlation between features that are not essential for recognition.
We then study the problem of strengthening neural architecture search in OoD scenarios.
arXiv Detail & Related papers (2024-10-25T20:50:32Z) - Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Unsupervised Discovery of Interpretable Directions in h-space of
Pre-trained Diffusion Models [63.1637853118899]
We propose the first unsupervised and learning-based method to identify interpretable directions in h-space of pre-trained diffusion models.
We employ a shift control module that works on h-space of pre-trained diffusion models to manipulate a sample into a shifted version of itself.
By jointly optimizing them, the model will spontaneously discover disentangled and interpretable directions.
arXiv Detail & Related papers (2023-10-15T18:44:30Z) - Elastic Interaction Energy-Based Generative Model: Approximation in
Feature Space [14.783344918500813]
We propose a novel approach to generative modeling using a loss function based on elastic interaction energy (EIE)
The utilization of the EIE-based metric presents several advantages, including its long range property that enables consideration of global information in the distribution.
Experimental results on popular datasets, such as MNIST, FashionMNIST, CIFAR-10, and CelebA, demonstrate that our EIEG GAN model can mitigate mode collapse, enhance stability, and improve model performance.
arXiv Detail & Related papers (2023-03-19T03:39:31Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Your GAN is Secretly an Energy-based Model and You Should use
Discriminator Driven Latent Sampling [106.68533003806276]
We show that sampling in latent space can be achieved by sampling in latent space according to an energy-based model induced by the sum of the latent prior log-density and the discriminator output score.
We show that Discriminator Driven Latent Sampling(DDLS) is highly efficient compared to previous methods which work in the high-dimensional pixel space.
arXiv Detail & Related papers (2020-03-12T23:33:50Z) - Simple and Effective Prevention of Mode Collapse in Deep One-Class
Classification [93.2334223970488]
We propose two regularizers to prevent hypersphere collapse in deep SVDD.
The first regularizer is based on injecting random noise via the standard cross-entropy loss.
The second regularizer penalizes the minibatch variance when it becomes too small.
arXiv Detail & Related papers (2020-01-24T03:44:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.