Scalable Balanced Training of Conditional Generative Adversarial Neural
Networks on Image Data
- URL: http://arxiv.org/abs/2102.10485v1
- Date: Sun, 21 Feb 2021 00:48:19 GMT
- Title: Scalable Balanced Training of Conditional Generative Adversarial Neural
Networks on Image Data
- Authors: Massimiliano Lupo Pasini, Vittorio Gabbi, Junqi Yin, Simona Perotto,
Nouamane Laanait
- Abstract summary: We propose a distributed approach to train deep convolutional generative adversarial neural network (DC-CGANs) models.
Our method reduces the imbalance between generator and discriminator by partitioning the training data according to data labels.
Performance is assessed in terms of inception score and image quality on MNIST, CIFAR10, CIFAR100, and ImageNet1k datasets.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a distributed approach to train deep convolutional generative
adversarial neural network (DC-CGANs) models. Our method reduces the imbalance
between generator and discriminator by partitioning the training data according
to data labels, and enhances scalability by performing a parallel training
where multiple generators are concurrently trained, each one of them focusing
on a single data label. Performance is assessed in terms of inception score and
image quality on MNIST, CIFAR10, CIFAR100, and ImageNet1k datasets, showing a
significant improvement in comparison to state-of-the-art techniques to
training DC-CGANs. Weak scaling is attained on all the four datasets using up
to 1,000 processes and 2,000 NVIDIA V100 GPUs on the OLCF supercomputer Summit.
Related papers
- DataDAM: Efficient Dataset Distillation with Attention Matching [15.300968899043498]
Researchers have long tried to minimize training costs in deep learning by maintaining strong generalization across diverse datasets.
Emerging research on dataset aims to reduce training costs by creating a small synthetic set that contains the information of a larger real dataset.
However, the synthetic data generated by previous methods are not guaranteed to distribute and discriminate as well as the original training data.
arXiv Detail & Related papers (2023-09-29T19:07:48Z) - Dataset Quantization [72.61936019738076]
We present dataset quantization (DQ), a new framework to compress large-scale datasets into small subsets.
DQ is the first method that can successfully distill large-scale datasets such as ImageNet-1k with a state-of-the-art compression ratio.
arXiv Detail & Related papers (2023-08-21T07:24:29Z) - Training on Thin Air: Improve Image Classification with Generated Data [28.96941414724037]
Diffusion Inversion is a simple yet effective method to generate diverse, high-quality training data for image classification.
Our approach captures the original data distribution and ensures data coverage by inverting images to the latent space of Stable Diffusion.
We identify three key components that allow our generated images to successfully supplant the original dataset.
arXiv Detail & Related papers (2023-05-24T16:33:02Z) - Stable Parallel Training of Wasserstein Conditional Generative
Adversarial Neural Networks [0.0]
We propose a stable, parallel approach to train Conditional Generative Adversarial Neural Networks (W-CGANs) under the constraint of a fixed computational budget.
We illustrate the approach on the CIFAR10, CIFAR100, and ImageNet1k datasets.
Performance is assessed in terms of scalability and final accuracy within a limited fixed computational time and computational resources.
arXiv Detail & Related papers (2022-07-25T16:30:40Z) - Feature transforms for image data augmentation [74.12025519234153]
In image classification, many augmentation approaches utilize simple image manipulation algorithms.
In this work, we build ensembles on the data level by adding images generated by combining fourteen augmentation approaches.
Pretrained ResNet50 networks are finetuned on training sets that include images derived from each augmentation method.
arXiv Detail & Related papers (2022-01-24T14:12:29Z) - Jigsaw Clustering for Unsupervised Visual Representation Learning [68.09280490213399]
We propose a new jigsaw clustering pretext task in this paper.
Our method makes use of information from both intra- and inter-images.
It is even comparable to the contrastive learning methods when only half of training batches are used.
arXiv Detail & Related papers (2021-04-01T08:09:26Z) - Ultra-Data-Efficient GAN Training: Drawing A Lottery Ticket First, Then
Training It Toughly [114.81028176850404]
Training generative adversarial networks (GANs) with limited data generally results in deteriorated performance and collapsed models.
We decompose the data-hungry GAN training into two sequential sub-problems.
Such a coordinated framework enables us to focus on lower-complexity and more data-efficient sub-problems.
arXiv Detail & Related papers (2021-02-28T05:20:29Z) - Quasi-Global Momentum: Accelerating Decentralized Deep Learning on
Heterogeneous Data [77.88594632644347]
Decentralized training of deep learning models is a key element for enabling data privacy and on-device learning over networks.
In realistic learning scenarios, the presence of heterogeneity across different clients' local datasets poses an optimization challenge.
We propose a novel momentum-based method to mitigate this decentralized training difficulty.
arXiv Detail & Related papers (2021-02-09T11:27:14Z) - Towards Faster and Stabilized GAN Training for High-fidelity Few-shot
Image Synthesis [21.40315235087551]
We propose a light-weight GAN structure that gains superior quality on 1024*1024 resolution.
We show our model's superior performance compared to the state-of-the-art StyleGAN2, when data and computing budget are limited.
arXiv Detail & Related papers (2021-01-12T22:02:54Z) - Differentiable Augmentation for Data-Efficient GAN Training [48.920992130257595]
We propose DiffAugment, a simple method that improves the data efficiency of GANs by imposing various types of differentiable augmentations on both real and fake samples.
Our method can generate high-fidelity images using only 100 images without pre-training, while being on par with existing transfer learning algorithms.
arXiv Detail & Related papers (2020-06-18T17:59:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.