One Model to Reconstruct Them All: A Novel Way to Use the Stochastic
Noise in StyleGAN
- URL: http://arxiv.org/abs/2010.11113v1
- Date: Wed, 21 Oct 2020 16:24:07 GMT
- Title: One Model to Reconstruct Them All: A Novel Way to Use the Stochastic
Noise in StyleGAN
- Authors: Christian Bartz, Joseph Bethge, Haojin Yang, Christoph Meinel
- Abstract summary: We present a novel StyleGAN-based autoencoder architecture, which can reconstruct images with very high quality across several data domains.
Our proposed architecture can handle up to 40 images per second on a single GPU, which is approximately 28x faster than previous approaches.
- Score: 10.810541849249821
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative Adversarial Networks (GANs) have achieved state-of-the-art
performance for several image generation and manipulation tasks. Different
works have improved the limited understanding of the latent space of GANs by
embedding images into specific GAN architectures to reconstruct the original
images. We present a novel StyleGAN-based autoencoder architecture, which can
reconstruct images with very high quality across several data domains. We
demonstrate a previously unknown grade of generalizablility by training the
encoder and decoder independently and on different datasets. Furthermore, we
provide new insights about the significance and capabilities of noise inputs of
the well-known StyleGAN architecture. Our proposed architecture can handle up
to 40 images per second on a single GPU, which is approximately 28x faster than
previous approaches. Finally, our model also shows promising results, when
compared to the state-of-the-art on the image denoising task, although it was
not explicitly designed for this task.
Related papers
- EGAIN: Extended GAn INversion [5.602947425285195]
Generative Adversarial Networks (GANs) have witnessed significant advances in recent years.
Recent GANs have proven to encode features in a disentangled latent space.
GAN inversion opens the door for the manipulation of facial semantics of real face images.
arXiv Detail & Related papers (2023-12-22T23:25:17Z) - Unlocking Pre-trained Image Backbones for Semantic Image Synthesis [29.688029979801577]
We propose a new class of GAN discriminators for semantic image synthesis that generates highly realistic images.
Our model, which we dub DP-SIMS, achieves state-of-the-art results in terms of image quality and consistency with the input label maps on ADE-20K, COCO-Stuff, and Cityscapes.
arXiv Detail & Related papers (2023-12-20T09:39:19Z) - Practical Blind Image Denoising via Swin-Conv-UNet and Data Synthesis [148.16279746287452]
We propose a swin-conv block to incorporate the local modeling ability of residual convolutional layer and non-local modeling ability of swin transformer block.
For the training data synthesis, we design a practical noise degradation model which takes into consideration different kinds of noise.
Experiments on AGWN removal and real image denoising demonstrate that the new network architecture design achieves state-of-the-art performance.
arXiv Detail & Related papers (2022-03-24T18:11:31Z) - InvGAN: Invertible GANs [88.58338626299837]
InvGAN, short for Invertible GAN, successfully embeds real images to the latent space of a high quality generative model.
This allows us to perform image inpainting, merging, and online data augmentation.
arXiv Detail & Related papers (2021-12-08T21:39:00Z) - Self-Denoising Neural Networks for Few Shot Learning [66.38505903102373]
We present a new training scheme that adds noise at multiple stages of an existing neural architecture while simultaneously learning to be robust to this added noise.
This architecture, which we call a Self-Denoising Neural Network (SDNN), can be applied easily to most modern convolutional neural architectures.
arXiv Detail & Related papers (2021-10-26T03:28:36Z) - Generating Diverse Structure for Image Inpainting With Hierarchical
VQ-VAE [74.29384873537587]
We propose a two-stage model for diverse inpainting, where the first stage generates multiple coarse results each of which has a different structure, and the second stage refines each coarse result separately by augmenting texture.
Experimental results on CelebA-HQ, Places2, and ImageNet datasets show that our method not only enhances the diversity of the inpainting solutions but also improves the visual quality of the generated multiple images.
arXiv Detail & Related papers (2021-03-18T05:10:49Z) - Multi-Stage Progressive Image Restoration [167.6852235432918]
We propose a novel synergistic design that can optimally balance these competing goals.
Our main proposal is a multi-stage architecture, that progressively learns restoration functions for the degraded inputs.
The resulting tightly interlinked multi-stage architecture, named as MPRNet, delivers strong performance gains on ten datasets.
arXiv Detail & Related papers (2021-02-04T18:57:07Z) - Spectrally Consistent UNet for High Fidelity Image Transformations [5.494315657902533]
Convolutional Neural Networks (CNNs) are the current de-facto models used for many imaging tasks.
In this work, a method for assessing the structural biases of UNets and the effects these have on the outputs is presented.
A new upsampling module is proposed, based on a novel use of the Guided Image Filter, that provides spectrally consistent outputs.
arXiv Detail & Related papers (2020-04-22T17:04:02Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z) - GAN Compression: Efficient Architectures for Interactive Conditional
GANs [45.012173624111185]
Recent Conditional Generative Adversarial Networks (cGANs) are 1-2 orders of magnitude more compute-intensive than modern recognition CNNs.
We propose a general-purpose compression framework for reducing the inference time and model size of the generator in cGANs.
arXiv Detail & Related papers (2020-03-19T17:59:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.