CoroNetGAN: Controlled Pruning of GANs via Hypernetworks
- URL: http://arxiv.org/abs/2403.08261v1
- Date: Wed, 13 Mar 2024 05:24:28 GMT
- Title: CoroNetGAN: Controlled Pruning of GANs via Hypernetworks
- Authors: Aman Kumar, Khushboo Anand, Shubham Mandloi, Ashutosh Mishra, Avinash
Thakur, Neeraj Kasera, Prathosh A P
- Abstract summary: We propose CoroNet-GAN for compressing GAN using the combined strength of differentiable pruning method via hypernetworks.
Our approach succeeds to outperform the baselines on Zebra-to-Horse and Summer-to-Winter achieving the best FID score of 32.3 and 72.3 respectively.
- Score: 5.765950477682605
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generative Adversarial Networks (GANs) have proven to exhibit remarkable
performance and are widely used across many generative computer vision
applications. However, the unprecedented demand for the deployment of GANs on
resource-constrained edge devices still poses a challenge due to huge number of
parameters involved in the generation process. This has led to focused
attention on the area of compressing GANs. Most of the existing works use
knowledge distillation with the overhead of teacher dependency. Moreover, there
is no ability to control the degree of compression in these methods. Hence, we
propose CoroNet-GAN for compressing GAN using the combined strength of
differentiable pruning method via hypernetworks. The proposed method provides
the advantage of performing controllable compression while training along with
reducing training time by a substantial factor. Experiments have been done on
various conditional GAN architectures (Pix2Pix and CycleGAN) to signify the
effectiveness of our approach on multiple benchmark datasets such as
Edges-to-Shoes, Horse-to-Zebra and Summer-to-Winter. The results obtained
illustrate that our approach succeeds to outperform the baselines on
Zebra-to-Horse and Summer-to-Winter achieving the best FID score of 32.3 and
72.3 respectively, yielding high-fidelity images across all the datasets.
Additionally, our approach also outperforms the state-of-the-art methods in
achieving better inference time on various smart-phone chipsets and data-types
making it a feasible solution for deployment on edge devices.
Related papers
- FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression [55.992528247880685]
Decentralized training faces significant challenges regarding system design and efficiency.
We present FusionLLM, a decentralized training system designed and implemented for training large deep neural networks (DNNs)
We show that our system and method can achieve 1.45 - 9.39x speedup compared to baseline methods while ensuring convergence.
arXiv Detail & Related papers (2024-10-16T16:13:19Z) - Nickel and Diming Your GAN: A Dual-Method Approach to Enhancing GAN Efficiency via Knowledge Distillation [8.330133104807759]
We propose two novel methodologies for compressing generative adversarial networks (GANs) in resource-constrained environments.
DiME and NICKEL achieve FID scores of 10.45 and 15.93 at compression rates of 95.73% and 98.92%, respectively.
Remarkably, our methods sustain generative quality even at an extreme compression rate of 99.69%, surpassing the previous state-of-the-art performance by a large margin.
arXiv Detail & Related papers (2024-05-19T17:09:43Z) - DuDGAN: Improving Class-Conditional GANs via Dual-Diffusion [2.458437232470188]
Class-conditional image generation using generative adversarial networks (GANs) has been investigated through various techniques.
We propose a novel approach for class-conditional image generation using GANs called DuDGAN, which incorporates a dual diffusion-based noise injection process.
Our method outperforms state-of-the-art conditional GAN models for image generation in terms of performance.
arXiv Detail & Related papers (2023-05-24T07:59:44Z) - Exploring Content Relationships for Distilling Efficient GANs [69.86835014810714]
This paper proposes a content relationship distillation (CRD) to tackle the over- parameterized generative adversarial networks (GANs)
In contrast to traditional instance-level distillation, we design a novel GAN compression oriented knowledge by slicing the contents of teacher outputs into multiple fine-grained granularities.
Built upon our proposed content-level distillation, we also deploy an online teacher discriminator, which keeps updating when co-trained with the teacher generator and keeps freezing when co-trained with the student generator for better adversarial training.
arXiv Detail & Related papers (2022-12-21T15:38:12Z) - GraphLearner: Graph Node Clustering with Fully Learnable Augmentation [76.63963385662426]
Contrastive deep graph clustering (CDGC) leverages the power of contrastive learning to group nodes into different clusters.
We propose a Graph Node Clustering with Fully Learnable Augmentation, termed GraphLearner.
It introduces learnable augmentors to generate high-quality and task-specific augmented samples for CDGC.
arXiv Detail & Related papers (2022-12-07T10:19:39Z) - DRAGON: Decentralized Fault Tolerance in Edge Federations [13.864161788250856]
We propose a novel memory-efficient deep learning based model, namely generative optimization networks (GON)
GONs use a single network to both discriminate input and generate samples, significantly reducing their memory footprint.
We propose a decentralized fault-tolerance method called DRAGON that runs simulations to quickly predict and optimize the performance of the edge federation.
arXiv Detail & Related papers (2022-08-16T10:40:28Z) - Hardware-Efficient Deconvolution-Based GAN for Edge Computing [1.5229257192293197]
Generative Adversarial Networks (GAN) are cutting-edge algorithms for generating new data samples based on the learned data distribution.
We proposed an HW/SW co-design approach for training quantized deconvolution GAN (QDCGAN) implemented on FPGA using a scalable streaming dataflow architecture.
Various precisions, datasets, and network scalability were analyzed for low-power inference on resource-constrained platforms.
arXiv Detail & Related papers (2022-01-18T11:16:59Z) - Learning Efficient GANs for Image Translation via Differentiable Masks
and co-Attention Distillation [130.30465659190773]
Generative Adversarial Networks (GANs) have been widely-used in image translation, but their high computation and storage costs impede the deployment on mobile devices.
We introduce a novel GAN compression method, termed DMAD, by proposing a Differentiable Mask and a co-Attention Distillation.
Experiments show DMAD can reduce the Multiply Accumulate Operations (MACs) of CycleGAN by 13x and that of Pix2Pix by 4x while retaining a comparable performance against the full model.
arXiv Detail & Related papers (2020-11-17T02:39:19Z) - Fitting the Search Space of Weight-sharing NAS with Graph Convolutional
Networks [100.14670789581811]
We train a graph convolutional network to fit the performance of sampled sub-networks.
With this strategy, we achieve a higher rank correlation coefficient in the selected set of candidates.
arXiv Detail & Related papers (2020-04-17T19:12:39Z) - Freeze the Discriminator: a Simple Baseline for Fine-Tuning GANs [104.85633684716296]
We show that simple fine-tuning of GANs with frozen lower layers of the discriminator performs surprisingly well.
This simple baseline, FreezeD, significantly outperforms previous techniques used in both unconditional and conditional GANs.
arXiv Detail & Related papers (2020-02-25T15:30:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.