Enhanced Balancing GAN: Minority-class Image Generation
- URL: http://arxiv.org/abs/2011.00189v1
- Date: Sat, 31 Oct 2020 05:03:47 GMT
- Title: Enhanced Balancing GAN: Minority-class Image Generation
- Authors: Gaofeng Huang and Amir H. Jafari
- Abstract summary: Generative adversarial networks (GANs) are one of the most powerful generative models.
Balancing GAN (BAGAN) is proposed to mitigate this problem, but it is unstable when images in different classes look similar.
In this work, we propose a supervised autoencoder with an intermediate embedding model to disperse the labeled latent vectors.
Our proposed model overcomes the unstable issue in original BAGAN and converges faster to high quality generations.
- Score: 0.7310043452300734
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative adversarial networks (GANs) are one of the most powerful
generative models, but always require a large and balanced dataset to train.
Traditional GANs are not applicable to generate minority-class images in a
highly imbalanced dataset. Balancing GAN (BAGAN) is proposed to mitigate this
problem, but it is unstable when images in different classes look similar, e.g.
flowers and cells. In this work, we propose a supervised autoencoder with an
intermediate embedding model to disperse the labeled latent vectors. With the
improved autoencoder initialization, we also build an architecture of BAGAN
with gradient penalty (BAGAN-GP). Our proposed model overcomes the unstable
issue in original BAGAN and converges faster to high quality generations. Our
model achieves high performance on the imbalanced scale-down version of MNIST
Fashion, CIFAR-10, and one small-scale medical image dataset.
Related papers
- Stabilize the Latent Space for Image Autoregressive Modeling: A Unified Perspective [52.778766190479374]
Latent-based image generative models have achieved notable success in image generation tasks.
Despite sharing the same latent space, autoregressive models significantly lag behind LDMs and MIMs in image generation.
We propose a simple but effective discrete image tokenizer to stabilize the latent space for image generative modeling.
arXiv Detail & Related papers (2024-10-16T12:13:17Z) - Model Inversion Attacks Through Target-Specific Conditional Diffusion Models [54.69008212790426]
Model inversion attacks (MIAs) aim to reconstruct private images from a target classifier's training set, thereby raising privacy concerns in AI applications.
Previous GAN-based MIAs tend to suffer from inferior generative fidelity due to GAN's inherent flaws and biased optimization within latent space.
We propose Diffusion-based Model Inversion (Diff-MI) attacks to alleviate these issues.
arXiv Detail & Related papers (2024-07-16T06:38:49Z) - Fair GANs through model rebalancing for extremely imbalanced class
distributions [5.463417677777276]
We present an approach to construct an unbiased generative adversarial network (GAN) from an existing biased GAN.
We show results for the StyleGAN2 models while training on the Flickr Faces High Quality (FFHQ) dataset for racial fairness.
We further validate our approach by applying it to an imbalanced CIFAR10 dataset which is also twice as large.
arXiv Detail & Related papers (2023-08-16T19:20:06Z) - LD-GAN: Low-Dimensional Generative Adversarial Network for Spectral
Image Generation with Variance Regularization [72.4394510913927]
Deep learning methods are state-of-the-art for spectral image (SI) computational tasks.
GANs enable diverse augmentation by learning and sampling from the data distribution.
GAN-based SI generation is challenging since the high-dimensionality nature of this kind of data hinders the convergence of the GAN training yielding to suboptimal generation.
We propose a statistical regularization to control the low-dimensional representation variance for the autoencoder training and to achieve high diversity of samples generated with the GAN.
arXiv Detail & Related papers (2023-04-29T00:25:02Z) - Traditional Classification Neural Networks are Good Generators: They are
Competitive with DDPMs and GANs [104.72108627191041]
We show that conventional neural network classifiers can generate high-quality images comparable to state-of-the-art generative models.
We propose a mask-based reconstruction module to make semantic gradients-aware to synthesize plausible images.
We show that our method is also applicable to text-to-image generation by regarding image-text foundation models.
arXiv Detail & Related papers (2022-11-27T11:25:35Z) - How to train your draGAN: A task oriented solution to imbalanced
classification [15.893327571516016]
This paper proposes a unique, performance-oriented, data-generating strategy that utilizes a new architecture, coined draGAN.
The samples are generated with the objective of optimizing the classification model's performance, rather than similarity to the real data.
Empirically we show the superiority of draGAN, but also highlight some of its shortcomings.
arXiv Detail & Related papers (2022-11-18T07:37:34Z) - FewGAN: Generating from the Joint Distribution of a Few Images [95.6635227371479]
We introduce FewGAN, a generative model for generating novel, high-quality and diverse images.
FewGAN is a hierarchical patch-GAN that applies quantization at the first coarse scale, followed by a pyramid of residual fully convolutional GANs at finer scales.
In an extensive set of experiments, it is shown that FewGAN outperforms baselines both quantitatively and qualitatively.
arXiv Detail & Related papers (2022-07-18T07:11:28Z) - Conditional Variational Autoencoder with Balanced Pre-training for
Generative Adversarial Networks [11.46883762268061]
Class imbalance occurs in many real-world applications, including image classification, where the number of images in each class differs significantly.
With imbalanced data, the generative adversarial networks (GANs) leans to majority class samples.
We propose a novel Variational Autoencoder with Balanced Pre-training for Geneversarative Adrial Networks (CAPGAN) as an augmentation tool to generate realistic synthetic images.
arXiv Detail & Related papers (2022-01-13T06:52:58Z) - InvGAN: Invertible GANs [88.58338626299837]
InvGAN, short for Invertible GAN, successfully embeds real images to the latent space of a high quality generative model.
This allows us to perform image inpainting, merging, and online data augmentation.
arXiv Detail & Related papers (2021-12-08T21:39:00Z) - Fast and Robust Cascade Model for Multiple Degradation Single Image
Super-Resolution [2.1574781022415364]
Single Image Super-Resolution (SISR) is one of the low-level computer vision problems that has received increased attention in the last few years.
Here, we propose a new formulation of the Convolutional Neural Network (CNN) cascade model.
A new densely connected CNN-architecture is proposed where the output of each sub- module is restricted using some external knowledge.
arXiv Detail & Related papers (2020-11-16T18:59:49Z) - TinyGAN: Distilling BigGAN for Conditional Image Generation [2.8072597424460466]
BigGAN has significantly improved the quality of image generation on ImageNet, but it requires a huge model, making it hard to deploy on resource-constrained devices.
We propose a black-box knowledge distillation framework for compressing GANs, which highlights a stable and efficient training process.
Given BigGAN as the teacher network, we manage to train a much smaller student network to mimic its functionality, achieving competitive performance on Inception and FID scores with the generator having $16times$ fewer parameters.
arXiv Detail & Related papers (2020-09-29T07:33:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.