Improving the Speed and Quality of GAN by Adversarial Training
- URL: http://arxiv.org/abs/2008.03364v1
- Date: Fri, 7 Aug 2020 20:21:31 GMT
- Title: Improving the Speed and Quality of GAN by Adversarial Training
- Authors: Jiachen Zhong, Xuanqing Liu, Cho-Jui Hsieh
- Abstract summary: We develop FastGAN to improve the speed and quality of GAN training based on the adversarial training technique.
Our training algorithm brings ImageNet training to the broader public by requiring 2-4 GPUs.
- Score: 87.70013107142142
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative adversarial networks (GAN) have shown remarkable results in image
generation tasks. High fidelity class-conditional GAN methods often rely on
stabilization techniques by constraining the global Lipschitz continuity. Such
regularization leads to less expressive models and slower convergence speed;
other techniques, such as the large batch training, require unconventional
computing power and are not widely accessible. In this paper, we develop an
efficient algorithm, namely FastGAN (Free AdverSarial Training), to improve the
speed and quality of GAN training based on the adversarial training technique.
We benchmark our method on CIFAR10, a subset of ImageNet, and the full ImageNet
datasets. We choose strong baselines such as SNGAN and SAGAN; the results
demonstrate that our training algorithm can achieve better generation quality
(in terms of the Inception score and Frechet Inception distance) with less
overall training time. Most notably, our training algorithm brings ImageNet
training to the broader public by requiring 2-4 GPUs.
Related papers
- ParaGAN: A Scalable Distributed Training Framework for Generative Adversarial Networks [8.736828068085263]
ParaGAN is a scalable distributed GAN training framework that leverages asynchronous training and asymmetric optimization policy to accelerate GAN training.
With ParaGAN, we reduce the training time of BigGAN from 15 days to 14 hours while achieving 91% scaling efficiency.
ParaGAN enables unprecedented high-resolution image generation using BigGAN.
arXiv Detail & Related papers (2024-11-06T15:40:46Z) - Auto-Train-Once: Controller Network Guided Automatic Network Pruning from Scratch [72.26822499434446]
Auto-Train-Once (ATO) is an innovative network pruning algorithm designed to automatically reduce the computational and storage costs of DNNs.
We provide a comprehensive convergence analysis as well as extensive experiments, and the results show that our approach achieves state-of-the-art performance across various model architectures.
arXiv Detail & Related papers (2024-03-21T02:33:37Z) - Always-Sparse Training by Growing Connections with Guided Stochastic
Exploration [46.4179239171213]
We propose an efficient always-sparse training algorithm with excellent scaling to larger and sparser models.
We evaluate our method on CIFAR-10/100 and ImageNet using VGG, and ViT models, and compare it against a range of sparsification methods.
arXiv Detail & Related papers (2024-01-12T21:32:04Z) - E$^{2}$GAN: Efficient Training of Efficient GANs for Image-to-Image Translation [69.72194342962615]
We introduce and address a novel research direction: can the process of distilling GANs from diffusion models be made significantly more efficient?
First, we construct a base GAN model with generalized features, adaptable to different concepts through fine-tuning, eliminating the need for training from scratch.
Second, we identify crucial layers within the base GAN model and employ Low-Rank Adaptation (LoRA) with a simple yet effective rank search process, rather than fine-tuning the entire base model.
Third, we investigate the minimal amount of data necessary for fine-tuning, further reducing the overall training time.
arXiv Detail & Related papers (2024-01-11T18:59:14Z) - Efficient Diffusion Training via Min-SNR Weighting Strategy [78.5801305960993]
We treat the diffusion training as a multi-task learning problem and introduce a simple yet effective approach referred to as Min-SNR-$gamma$.
Our results demonstrate a significant improvement in converging speed, 3.4$times$ faster than previous weighting strategies.
It is also more effective, achieving a new record FID score of 2.06 on the ImageNet $256times256$ benchmark using smaller architectures than that employed in previous state-of-the-art.
arXiv Detail & Related papers (2023-03-16T17:59:56Z) - Time Efficient Training of Progressive Generative Adversarial Network
using Depthwise Separable Convolution and Super Resolution Generative
Adversarial Network [0.0]
We propose a novel pipeline that combines Progressive GAN with slight modifications and Super Resolution GAN.
Super Resolution GAN up samples low-resolution images to high-resolution images which can prove to be a useful resource to reduce the training time exponentially.
arXiv Detail & Related papers (2022-02-24T19:53:37Z) - Ultra-Data-Efficient GAN Training: Drawing A Lottery Ticket First, Then
Training It Toughly [114.81028176850404]
Training generative adversarial networks (GANs) with limited data generally results in deteriorated performance and collapsed models.
We decompose the data-hungry GAN training into two sequential sub-problems.
Such a coordinated framework enables us to focus on lower-complexity and more data-efficient sub-problems.
arXiv Detail & Related papers (2021-02-28T05:20:29Z) - Towards Faster and Stabilized GAN Training for High-fidelity Few-shot
Image Synthesis [21.40315235087551]
We propose a light-weight GAN structure that gains superior quality on 1024*1024 resolution.
We show our model's superior performance compared to the state-of-the-art StyleGAN2, when data and computing budget are limited.
arXiv Detail & Related papers (2021-01-12T22:02:54Z) - Improving GAN Training with Probability Ratio Clipping and Sample
Reweighting [145.5106274085799]
generative adversarial networks (GANs) often suffer from inferior performance due to unstable training.
We propose a new variational GAN training framework which enjoys superior training stability.
By plugging the training approach in diverse state-of-the-art GAN architectures, we obtain significantly improved performance over a range of tasks.
arXiv Detail & Related papers (2020-06-12T01:39:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.