Ultra-Data-Efficient GAN Training: Drawing A Lottery Ticket First, Then
Training It Toughly
- URL: http://arxiv.org/abs/2103.00397v1
- Date: Sun, 28 Feb 2021 05:20:29 GMT
- Title: Ultra-Data-Efficient GAN Training: Drawing A Lottery Ticket First, Then
Training It Toughly
- Authors: Tianlong Chen, Yu Cheng, Zhe Gan, Jingjing Liu, Zhangyang Wang
- Abstract summary: Training generative adversarial networks (GANs) with limited data generally results in deteriorated performance and collapsed models.
We decompose the data-hungry GAN training into two sequential sub-problems.
Such a coordinated framework enables us to focus on lower-complexity and more data-efficient sub-problems.
- Score: 114.81028176850404
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Training generative adversarial networks (GANs) with limited data generally
results in deteriorated performance and collapsed models. To conquer this
challenge, we are inspired by the latest observation of Kalibhat et al. (2020);
Chen et al.(2021d), that one can discover independently trainable and highly
sparse subnetworks (a.k.a., lottery tickets) from GANs. Treating this as an
inductive prior, we decompose the data-hungry GAN training into two sequential
sub-problems: (i) identifying the lottery ticket from the original GAN; then
(ii) training the found sparse subnetwork with aggressive data and feature
augmentations. Both sub-problems re-use the same small training set of real
images. Such a coordinated framework enables us to focus on lower-complexity
and more data-efficient sub-problems, effectively stabilizing training and
improving convergence. Comprehensive experiments endorse the effectiveness of
our proposed ultra-data-efficient training framework, across various GAN
architectures (SNGAN, BigGAN, and StyleGAN2) and diverse datasets (CIFAR-10,
CIFAR-100, Tiny-ImageNet, and ImageNet). Besides, our training framework also
displays powerful few-shot generalization ability, i.e., generating
high-fidelity images by training from scratch with just 100 real images,
without any pre-training. Codes are available at:
https://github.com/VITA-Group/Ultra-Data-Efficient-GAN-Training.
Related papers
- Improving GAN Training via Feature Space Shrinkage [69.98365478398593]
We propose AdaptiveMix, which shrinks regions of training data in the image representation space of the discriminator.
Considering it is intractable to directly bound feature space, we propose to construct hard samples and narrow down the feature distance between hard and easy samples.
The evaluation results demonstrate that our AdaptiveMix can facilitate the training of GANs and effectively improve the image quality of generated samples.
arXiv Detail & Related papers (2023-03-02T20:22:24Z) - Sparsity Winning Twice: Better Robust Generalization from More Efficient
Training [94.92954973680914]
We introduce two alternatives for sparse adversarial training: (i) static sparsity and (ii) dynamic sparsity.
We find both methods to yield win-win: substantially shrinking the robust generalization gap and alleviating the robust overfitting.
Our approaches can be combined with existing regularizers, establishing new state-of-the-art results in adversarial training.
arXiv Detail & Related papers (2022-02-20T15:52:08Z) - Regularizing Generative Adversarial Networks under Limited Data [88.57330330305535]
This work proposes a regularization approach for training robust GAN models on limited data.
We show a connection between the regularized loss and an f-divergence called LeCam-divergence, which we find is more robust under limited training data.
arXiv Detail & Related papers (2021-04-07T17:59:06Z) - Scalable Balanced Training of Conditional Generative Adversarial Neural
Networks on Image Data [0.0]
We propose a distributed approach to train deep convolutional generative adversarial neural network (DC-CGANs) models.
Our method reduces the imbalance between generator and discriminator by partitioning the training data according to data labels.
Performance is assessed in terms of inception score and image quality on MNIST, CIFAR10, CIFAR100, and ImageNet1k datasets.
arXiv Detail & Related papers (2021-02-21T00:48:19Z) - Towards Faster and Stabilized GAN Training for High-fidelity Few-shot
Image Synthesis [21.40315235087551]
We propose a light-weight GAN structure that gains superior quality on 1024*1024 resolution.
We show our model's superior performance compared to the state-of-the-art StyleGAN2, when data and computing budget are limited.
arXiv Detail & Related papers (2021-01-12T22:02:54Z) - Multiclass non-Adversarial Image Synthesis, with Application to
Classification from Very Small Sample [6.243995448840211]
We present a novel non-adversarial generative method - Clustered Optimization of LAtent space (COLA)
In the full data regime, our method is capable of generating diverse multi-class images with no supervision.
In the small-data regime, where only a small sample of labeled images is available for training with no access to additional unlabeled data, our results surpass state-of-the-art GAN models trained on the same amount of data.
arXiv Detail & Related papers (2020-11-25T18:47:27Z) - Improving the Speed and Quality of GAN by Adversarial Training [87.70013107142142]
We develop FastGAN to improve the speed and quality of GAN training based on the adversarial training technique.
Our training algorithm brings ImageNet training to the broader public by requiring 2-4 GPUs.
arXiv Detail & Related papers (2020-08-07T20:21:31Z) - Differentiable Augmentation for Data-Efficient GAN Training [48.920992130257595]
We propose DiffAugment, a simple method that improves the data efficiency of GANs by imposing various types of differentiable augmentations on both real and fake samples.
Our method can generate high-fidelity images using only 100 images without pre-training, while being on par with existing transfer learning algorithms.
arXiv Detail & Related papers (2020-06-18T17:59:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.