Annealing Genetic GAN for Minority Oversampling
- URL: http://arxiv.org/abs/2008.01967v1
- Date: Wed, 5 Aug 2020 07:19:47 GMT
- Title: Annealing Genetic GAN for Minority Oversampling
- Authors: Jingyu Hao and Chengjia Wang and Heye Zhang and Guang Yang
- Abstract summary: Generative Adversarial Networks (GANs) have shown some potentials to tackle class imbalance problems.
We propose an Annealing Genetic GAN (AGGAN) method, which aims to reproduce the distributions closest to the ones of the minority classes using only limited data samples.
- Score: 5.818339336603936
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The key to overcome class imbalance problems is to capture the distribution
of minority class accurately. Generative Adversarial Networks (GANs) have shown
some potentials to tackle class imbalance problems due to their capability of
reproducing data distributions given ample training data samples. However, the
scarce samples of one or more classes still pose a great challenge for GANs to
learn accurate distributions for the minority classes. In this work, we propose
an Annealing Genetic GAN (AGGAN) method, which aims to reproduce the
distributions closest to the ones of the minority classes using only limited
data samples. Our AGGAN renovates the training of GANs as an evolutionary
process that incorporates the mechanism of simulated annealing. In particular,
the generator uses different training strategies to generate multiple offspring
and retain the best. Then, we use the Metropolis criterion in the simulated
annealing to decide whether we should update the best offspring for the
generator. As the Metropolis criterion allows a certain chance to accept the
worse solutions, it enables our AGGAN steering away from the local optimum.
According to both theoretical analysis and experimental studies on multiple
imbalanced image datasets, we prove that the proposed training strategy can
enable our AGGAN to reproduce the distributions of minority classes from scarce
samples and provide an effective and robust solution for the class imbalance
problem.
Related papers
- Tackling Diverse Minorities in Imbalanced Classification [80.78227787608714]
Imbalanced datasets are commonly observed in various real-world applications, presenting significant challenges in training classifiers.
We propose generating synthetic samples iteratively by mixing data samples from both minority and majority classes.
We demonstrate the effectiveness of our proposed framework through extensive experiments conducted on seven publicly available benchmark datasets.
arXiv Detail & Related papers (2023-08-28T18:48:34Z) - Reweighted Mixup for Subpopulation Shift [63.1315456651771]
Subpopulation shift exists in many real-world applications, which refers to the training and test distributions that contain the same subpopulation groups but with different subpopulation proportions.
Importance reweighting is a classical and effective way to handle the subpopulation shift.
We propose a simple yet practical framework, called reweighted mixup, to mitigate the overfitting issue.
arXiv Detail & Related papers (2023-04-09T03:44:50Z) - Imbalanced Class Data Performance Evaluation and Improvement using Novel
Generative Adversarial Network-based Approach: SSG and GBO [0.0]
This study proposes two novel techniques: GAN-based Oversampling (GBO) and Support Vector Machine-SMOTE-GAN (SSG)
The preliminary computational result shows that SSG and GBO performed better on the expanded imbalanced eight benchmark datasets than the original SMOTE.
arXiv Detail & Related papers (2022-10-23T22:17:54Z) - Learning to Re-weight Examples with Optimal Transport for Imbalanced
Classification [74.62203971625173]
Imbalanced data pose challenges for deep learning based classification models.
One of the most widely-used approaches for tackling imbalanced data is re-weighting.
We propose a novel re-weighting method based on optimal transport (OT) from a distributional point of view.
arXiv Detail & Related papers (2022-08-05T01:23:54Z) - Distributionally Robust Models with Parametric Likelihood Ratios [123.05074253513935]
Three simple ideas allow us to train models with DRO using a broader class of parametric likelihood ratios.
We find that models trained with the resulting parametric adversaries are consistently more robust to subpopulation shifts when compared to other DRO approaches.
arXiv Detail & Related papers (2022-04-13T12:43:12Z) - Deceive D: Adaptive Pseudo Augmentation for GAN Training with Limited
Data [125.7135706352493]
Generative adversarial networks (GANs) typically require ample data for training in order to synthesize high-fidelity images.
Recent studies have shown that training GANs with limited data remains formidable due to discriminator overfitting.
This paper introduces a novel strategy called Adaptive Pseudo Augmentation (APA) to encourage healthy competition between the generator and the discriminator.
arXiv Detail & Related papers (2021-11-12T18:13:45Z) - SMOTified-GAN for class imbalanced pattern classification problems [0.41998444721319217]
We propose a novel two-phase oversampling approach that has the synergy of SMOTE and GAN.
The experimental results prove the sample quality of minority class(es) has been improved in a variety of tested benchmark datasets.
arXiv Detail & Related papers (2021-08-06T06:14:05Z) - MineGAN++: Mining Generative Models for Efficient Knowledge Transfer to
Limited Data Domains [77.46963293257912]
We propose a novel knowledge transfer method for generative models based on mining the knowledge that is most beneficial to a specific target domain.
This is done using a miner network that identifies which part of the generative distribution of each pretrained GAN outputs samples closest to the target domain.
We show that the proposed method, called MineGAN, effectively transfers knowledge to domains with few target images, outperforming existing methods.
arXiv Detail & Related papers (2021-04-28T13:10:56Z) - Gamma distribution-based sampling for imbalanced data [6.85316573653194]
Imbalanced class distribution is a common problem in a number of fields including medical diagnostics, fraud detection, and others.
We propose a novel method for balancing the class distribution in data through intelligent resampling of the minority class instances.
arXiv Detail & Related papers (2020-09-22T06:39:13Z) - The Bures Metric for Generative Adversarial Networks [10.69910379275607]
Generative Adversarial Networks (GANs) are performant generative methods yielding high-quality samples.
We propose to match the real batch diversity to the fake batch diversity.
We observe that diversity matching reduces mode collapse substantially and has a positive effect on the sample quality.
arXiv Detail & Related papers (2020-06-16T12:04:41Z) - Imbalanced Data Learning by Minority Class Augmentation using Capsule
Adversarial Networks [31.073558420480964]
We propose a method to restore the balance in imbalanced images, by coalescing two concurrent methods.
In our model, generative and discriminative networks play a novel competitive game.
The coalescing of capsule-GAN is effective at recognizing highly overlapping classes with much fewer parameters compared with the convolutional-GAN.
arXiv Detail & Related papers (2020-04-05T12:36:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.