GAN Compression: Efficient Architectures for Interactive Conditional
GANs
- URL: http://arxiv.org/abs/2003.08936v4
- Date: Thu, 11 Nov 2021 03:45:16 GMT
- Title: GAN Compression: Efficient Architectures for Interactive Conditional
GANs
- Authors: Muyang Li, Ji Lin, Yaoyao Ding, Zhijian Liu, Jun-Yan Zhu, Song Han
- Abstract summary: Recent Conditional Generative Adversarial Networks (cGANs) are 1-2 orders of magnitude more compute-intensive than modern recognition CNNs.
We propose a general-purpose compression framework for reducing the inference time and model size of the generator in cGANs.
- Score: 45.012173624111185
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conditional Generative Adversarial Networks (cGANs) have enabled controllable
image synthesis for many vision and graphics applications. However, recent
cGANs are 1-2 orders of magnitude more compute-intensive than modern
recognition CNNs. For example, GauGAN consumes 281G MACs per image, compared to
0.44G MACs for MobileNet-v3, making it difficult for interactive deployment. In
this work, we propose a general-purpose compression framework for reducing the
inference time and model size of the generator in cGANs. Directly applying
existing compression methods yields poor performance due to the difficulty of
GAN training and the differences in generator architectures. We address these
challenges in two ways. First, to stabilize GAN training, we transfer knowledge
of multiple intermediate representations of the original model to its
compressed model and unify unpaired and paired learning. Second, instead of
reusing existing CNN designs, our method finds efficient architectures via
neural architecture search. To accelerate the search process, we decouple the
model training and search via weight sharing. Experiments demonstrate the
effectiveness of our method across different supervision settings, network
architectures, and learning methods. Without losing image quality, we reduce
the computation of CycleGAN by 21x, Pix2pix by 12x, MUNIT by 29x, and GauGAN by
9x, paving the way for interactive image synthesis.
Related papers
- Distance Weighted Trans Network for Image Completion [52.318730994423106]
We propose a new architecture that relies on Distance-based Weighted Transformer (DWT) to better understand the relationships between an image's components.
CNNs are used to augment the local texture information of coarse priors.
DWT blocks are used to recover certain coarse textures and coherent visual structures.
arXiv Detail & Related papers (2023-10-11T12:46:11Z) - InvGAN: Invertible GANs [88.58338626299837]
InvGAN, short for Invertible GAN, successfully embeds real images to the latent space of a high quality generative model.
This allows us to perform image inpainting, merging, and online data augmentation.
arXiv Detail & Related papers (2021-12-08T21:39:00Z) - FMD-cGAN: Fast Motion Deblurring using Conditional Generative
Adversarial Networks [26.878173373199786]
We present a Fast Motion Deblurring-Conditional Generative Adversarial Network (FMD-cGAN) that helps in blind motion deblurring of a single image.
FMD-cGAN delivers impressive structural similarity and visual appearance after deblurring an image.
arXiv Detail & Related papers (2021-11-30T14:30:44Z) - Learning Efficient GANs for Image Translation via Differentiable Masks
and co-Attention Distillation [130.30465659190773]
Generative Adversarial Networks (GANs) have been widely-used in image translation, but their high computation and storage costs impede the deployment on mobile devices.
We introduce a novel GAN compression method, termed DMAD, by proposing a Differentiable Mask and a co-Attention Distillation.
Experiments show DMAD can reduce the Multiply Accumulate Operations (MACs) of CycleGAN by 13x and that of Pix2Pix by 4x while retaining a comparable performance against the full model.
arXiv Detail & Related papers (2020-11-17T02:39:19Z) - Efficient texture-aware multi-GAN for image inpainting [5.33024001730262]
Recent GAN-based (Generative adversarial networks) inpainting methods show remarkable improvements.
We propose a multi-GAN architecture improving both the performance and rendering efficiency.
arXiv Detail & Related papers (2020-09-30T14:58:03Z) - GAN Slimming: All-in-One GAN Compression by A Unified Optimization
Framework [94.26938614206689]
We propose the first unified optimization framework combining multiple compression means for GAN compression, dubbed GAN Slimming.
We apply GS to compress CartoonGAN, a state-of-the-art style transfer network, by up to 47 times, with minimal visual quality degradation.
arXiv Detail & Related papers (2020-08-25T14:39:42Z) - Fidelity-Controllable Extreme Image Compression with Generative
Adversarial Networks [10.036312061637764]
We propose a GAN-based image compression method working at extremely lows below 0.1bpp.
To address both of the drawbacks, our method adopts two-stage training and network.
The experimental results show that our model can reconstruct high quality images.
arXiv Detail & Related papers (2020-08-24T10:45:19Z) - InfoMax-GAN: Improved Adversarial Image Generation via Information
Maximization and Contrastive Learning [39.316605441868944]
Generative Adversarial Networks (GANs) are fundamental to many generative modelling applications.
We propose a principled framework to simultaneously mitigate two fundamental issues in GANs: catastrophic forgetting of the discriminator and mode collapse of the generator.
Our approach significantly stabilizes GAN training and improves GAN performance for image synthesis across five datasets.
arXiv Detail & Related papers (2020-07-09T06:56:11Z) - Freeze the Discriminator: a Simple Baseline for Fine-Tuning GANs [104.85633684716296]
We show that simple fine-tuning of GANs with frozen lower layers of the discriminator performs surprisingly well.
This simple baseline, FreezeD, significantly outperforms previous techniques used in both unconditional and conditional GANs.
arXiv Detail & Related papers (2020-02-25T15:30:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.