Discriminator-Cooperated Feature Map Distillation for GAN Compression
- URL: http://arxiv.org/abs/2212.14169v1
- Date: Thu, 29 Dec 2022 03:50:27 GMT
- Title: Discriminator-Cooperated Feature Map Distillation for GAN Compression
- Authors: Tie Hu, Mingbao Lin, Lizhou You, Fei Chao, Rongrong Ji
- Abstract summary: We present an inventive discriminator-cooperated distillation, abbreviated as DCD, towards refining better feature maps from the generator.
Our DCD shows superior results compared with existing GAN compression methods.
- Score: 69.86835014810714
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite excellent performance in image generation, Generative Adversarial
Networks (GANs) are notorious for its requirements of enormous storage and
intensive computation. As an awesome ''performance maker'', knowledge
distillation is demonstrated to be particularly efficacious in exploring
low-priced GANs. In this paper, we investigate the irreplaceability of teacher
discriminator and present an inventive discriminator-cooperated distillation,
abbreviated as DCD, towards refining better feature maps from the generator. In
contrast to conventional pixel-to-pixel match methods in feature map
distillation, our DCD utilizes teacher discriminator as a transformation to
drive intermediate results of the student generator to be perceptually close to
corresponding outputs of the teacher generator. Furthermore, in order to
mitigate mode collapse in GAN compression, we construct a collaborative
adversarial training paradigm where the teacher discriminator is from scratch
established to co-train with student generator in company with our DCD. Our DCD
shows superior results compared with existing GAN compression methods. For
instance, after reducing over 40x MACs and 80x parameters of CycleGAN, we well
decrease FID metric from 61.53 to 48.24 while the current SoTA method merely
has 51.92. This work's source code has been made accessible at
https://github.com/poopit/DCD-official.
Related papers
- SD-DiT: Unleashing the Power of Self-supervised Discrimination in Diffusion Transformer [102.39050180060913]
Diffusion Transformer (DiT) has emerged as the new trend of generative diffusion models on image generation.
Recent breakthroughs have been driven by mask strategy that significantly improves the training efficiency of DiT with additional intra-image contextual learning.
In this work, we address these limitations by novelly unleashing the self-supervised discrimination knowledge to boost DiT training.
arXiv Detail & Related papers (2024-03-25T17:59:35Z) - Exploring Content Relationships for Distilling Efficient GANs [69.86835014810714]
This paper proposes a content relationship distillation (CRD) to tackle the over- parameterized generative adversarial networks (GANs)
In contrast to traditional instance-level distillation, we design a novel GAN compression oriented knowledge by slicing the contents of teacher outputs into multiple fine-grained granularities.
Built upon our proposed content-level distillation, we also deploy an online teacher discriminator, which keeps updating when co-trained with the teacher generator and keeps freezing when co-trained with the student generator for better adversarial training.
arXiv Detail & Related papers (2022-12-21T15:38:12Z) - Revisiting Discriminator in GAN Compression: A Generator-discriminator
Cooperative Compression Scheme [65.5405625485559]
GAN compression aims to reduce tremendous computational overhead and memory usages when deploying GANs on resource-constrained edge devices.
In this work, we revisit the role of discriminator in GAN compression and design a novel generator-discriminator cooperative compression scheme for GAN compression, termed GCC.
arXiv Detail & Related papers (2021-10-27T13:54:55Z) - Online Multi-Granularity Distillation for GAN Compression [17.114017187236836]
Generative Adversarial Networks (GANs) have witnessed prevailing success in yielding outstanding images.
GANs are burdensome to deploy on resource-constrained devices due to ponderous computational costs and hulking memory usage.
We propose a novel online multi-granularity distillation scheme to obtain lightweight GANs.
arXiv Detail & Related papers (2021-08-16T05:49:50Z) - Learning Efficient GANs for Image Translation via Differentiable Masks
and co-Attention Distillation [130.30465659190773]
Generative Adversarial Networks (GANs) have been widely-used in image translation, but their high computation and storage costs impede the deployment on mobile devices.
We introduce a novel GAN compression method, termed DMAD, by proposing a Differentiable Mask and a co-Attention Distillation.
Experiments show DMAD can reduce the Multiply Accumulate Operations (MACs) of CycleGAN by 13x and that of Pix2Pix by 4x while retaining a comparable performance against the full model.
arXiv Detail & Related papers (2020-11-17T02:39:19Z) - Discriminator Contrastive Divergence: Semi-Amortized Generative Modeling
by Exploring Energy of the Discriminator [85.68825725223873]
Generative Adversarial Networks (GANs) have shown great promise in modeling high dimensional data.
We introduce the Discriminator Contrastive Divergence, which is well motivated by the property of WGAN's discriminator.
We demonstrate the benefits of significant improved generation on both synthetic data and several real-world image generation benchmarks.
arXiv Detail & Related papers (2020-04-05T01:50:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.