Stable Parallel Training of Wasserstein Conditional Generative
Adversarial Neural Networks
- URL: http://arxiv.org/abs/2207.12315v1
- Date: Mon, 25 Jul 2022 16:30:40 GMT
- Title: Stable Parallel Training of Wasserstein Conditional Generative
Adversarial Neural Networks
- Authors: Massimiliano Lupo Pasini, Junqi Yin
- Abstract summary: We propose a stable, parallel approach to train Conditional Generative Adversarial Neural Networks (W-CGANs) under the constraint of a fixed computational budget.
We illustrate the approach on the CIFAR10, CIFAR100, and ImageNet1k datasets.
Performance is assessed in terms of scalability and final accuracy within a limited fixed computational time and computational resources.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a stable, parallel approach to train Wasserstein Conditional
Generative Adversarial Neural Networks (W-CGANs) under the constraint of a
fixed computational budget. Differently from previous distributed GANs training
techniques, our approach avoids inter-process communications, reduces the risk
of mode collapse and enhances scalability by using multiple generators, each
one of them concurrently trained on a single data label. The use of the
Wasserstein metric also reduces the risk of cycling by stabilizing the training
of each generator. We illustrate the approach on the CIFAR10, CIFAR100, and
ImageNet1k datasets, three standard benchmark image datasets, maintaining the
original resolution of the images for each dataset. Performance is assessed in
terms of scalability and final accuracy within a limited fixed computational
time and computational resources. To measure accuracy, we use the inception
score, the Frechet inception distance, and image quality. An improvement in
inception score and Frechet inception distance is shown in comparison to
previous results obtained by performing the parallel approach on deep
convolutional conditional generative adversarial neural networks (DC-CGANs) as
well as an improvement of image quality of the new images created by the GANs
approach. Weak scaling is attained on both datasets using up to 2,000 NVIDIA
V100 GPUs on the OLCF supercomputer Summit.
Related papers
- TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals [58.865901821451295]
We present a novel two-stream feature fusion "Tensor-Convolution and Convolution-Transformer Network" (TCCT-Net) architecture.
To better learn the meaningful patterns in the temporal-spatial domain, we design a "CT" stream that integrates a hybrid convolutional-transformer.
In parallel, to efficiently extract rich patterns from the temporal-frequency domain, we introduce a "TC" stream that uses Continuous Wavelet Transform (CWT) to represent information in a 2D tensor form.
arXiv Detail & Related papers (2024-04-15T06:01:48Z) - Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - LD-GAN: Low-Dimensional Generative Adversarial Network for Spectral
Image Generation with Variance Regularization [72.4394510913927]
Deep learning methods are state-of-the-art for spectral image (SI) computational tasks.
GANs enable diverse augmentation by learning and sampling from the data distribution.
GAN-based SI generation is challenging since the high-dimensionality nature of this kind of data hinders the convergence of the GAN training yielding to suboptimal generation.
We propose a statistical regularization to control the low-dimensional representation variance for the autoencoder training and to achieve high diversity of samples generated with the GAN.
arXiv Detail & Related papers (2023-04-29T00:25:02Z) - Enhanced Sharp-GAN For Histopathology Image Synthesis [63.845552349914186]
Histopathology image synthesis aims to address the data shortage issue in training deep learning approaches for accurate cancer detection.
We propose a novel approach that enhances the quality of synthetic images by using nuclei topology and contour regularization.
The proposed approach outperforms Sharp-GAN in all four image quality metrics on two datasets.
arXiv Detail & Related papers (2023-01-24T17:54:01Z) - Curvature regularization for Non-line-of-sight Imaging from
Under-sampled Data [5.591221518341613]
Non-line-of-sight (NLOS) imaging aims to reconstruct the three-dimensional hidden scenes from the data measured in the line-of-sight.
We propose novel NLOS reconstruction models based on curvature regularization.
We evaluate the proposed algorithms on both synthetic and real datasets.
arXiv Detail & Related papers (2023-01-01T14:10:43Z) - Feature transforms for image data augmentation [74.12025519234153]
In image classification, many augmentation approaches utilize simple image manipulation algorithms.
In this work, we build ensembles on the data level by adding images generated by combining fourteen augmentation approaches.
Pretrained ResNet50 networks are finetuned on training sets that include images derived from each augmentation method.
arXiv Detail & Related papers (2022-01-24T14:12:29Z) - Deep Neural Networks Based Weight Approximation and Computation Reuse
for 2-D Image Classification [0.9507070656654631]
Deep Neural Networks (DNNs) are computationally and memory intensive.
This paper introduces a new method to improve DNNs performance by fusing approximate computing with data reuse techniques.
It is suitable for IoT edge devices as it reduces the memory size requirement as well as the number of needed memory accesses.
arXiv Detail & Related papers (2021-04-28T10:16:53Z) - Scalable Balanced Training of Conditional Generative Adversarial Neural
Networks on Image Data [0.0]
We propose a distributed approach to train deep convolutional generative adversarial neural network (DC-CGANs) models.
Our method reduces the imbalance between generator and discriminator by partitioning the training data according to data labels.
Performance is assessed in terms of inception score and image quality on MNIST, CIFAR10, CIFAR100, and ImageNet1k datasets.
arXiv Detail & Related papers (2021-02-21T00:48:19Z) - Sparse Signal Models for Data Augmentation in Deep Learning ATR [0.8999056386710496]
We propose a data augmentation approach to incorporate domain knowledge and improve the generalization power of a data-intensive learning algorithm.
We exploit the sparsity of the scattering centers in the spatial domain and the smoothly-varying structure of the scattering coefficients in the azimuthal domain to solve the ill-posed problem of over-parametrized model fitting.
arXiv Detail & Related papers (2020-12-16T21:46:33Z) - Locally Masked Convolution for Autoregressive Models [107.4635841204146]
LMConv is a simple modification to the standard 2D convolution that allows arbitrary masks to be applied to the weights at each location in the image.
We learn an ensemble of distribution estimators that share parameters but differ in generation order, achieving improved performance on whole-image density estimation.
arXiv Detail & Related papers (2020-06-22T17:59:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.