DeepI2I: Enabling Deep Hierarchical Image-to-Image Translation by
Transferring from GANs
- URL: http://arxiv.org/abs/2011.05867v1
- Date: Wed, 11 Nov 2020 16:03:03 GMT
- Title: DeepI2I: Enabling Deep Hierarchical Image-to-Image Translation by
Transferring from GANs
- Authors: Yaxing Wang, Lu Yu, Joost van de Weijer
- Abstract summary: Image-to-image translation suffers from inferior performance when translations between classes require large shape changes.
We propose a novel deep hierarchical Image-to-Image Translation method, called DeepI2I.
We demonstrate that transfer learning significantly improves the performance of I2I systems, especially for small datasets.
- Score: 43.33066765114446
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image-to-image translation has recently achieved remarkable results. But
despite current success, it suffers from inferior performance when translations
between classes require large shape changes. We attribute this to the
high-resolution bottlenecks which are used by current state-of-the-art
image-to-image methods. Therefore, in this work, we propose a novel deep
hierarchical Image-to-Image Translation method, called DeepI2I. We learn a
model by leveraging hierarchical features: (a) structural information contained
in the shallow layers and (b) semantic information extracted from the deep
layers. To enable the training of deep I2I models on small datasets, we propose
a novel transfer learning method, that transfers knowledge from pre-trained
GANs. Specifically, we leverage the discriminator of a pre-trained GANs (i.e.
BigGAN or StyleGAN) to initialize both the encoder and the discriminator and
the pre-trained generator to initialize the generator of our model. Applying
knowledge transfer leads to an alignment problem between the encoder and
generator. We introduce an adaptor network to address this. On many-class
image-to-image translation on three datasets (Animal faces, Birds, and Foods)
we decrease mFID by at least 35% when compared to the state-of-the-art.
Furthermore, we qualitatively and quantitatively demonstrate that transfer
learning significantly improves the performance of I2I systems, especially for
small datasets. Finally, we are the first to perform I2I translations for
domains with over 100 classes.
Related papers
- I2I-Galip: Unsupervised Medical Image Translation Using Generative Adversarial CLIP [30.506544165999564]
Unpaired image-to-image translation is a challenging task due to the absence of paired examples.
We propose a new image-to-image translation framework named Image-to-Image-Generative-Adversarial-CLIP (I2I-Galip)
arXiv Detail & Related papers (2024-09-19T01:44:50Z) - Additional Look into GAN-based Augmentation for Deep Learning COVID-19
Image Classification [57.1795052451257]
We study the dependence of the GAN-based augmentation performance on dataset size with a focus on small samples.
We train StyleGAN2-ADA with both sets and then, after validating the quality of generated images, we use trained GANs as one of the augmentations approaches in multi-class classification problems.
The GAN-based augmentation approach is found to be comparable with classical augmentation in the case of medium and large datasets but underperforms in the case of smaller datasets.
arXiv Detail & Related papers (2024-01-26T08:28:13Z) - Guided Image-to-Image Translation by Discriminator-Generator
Communication [71.86347329356244]
The goal of Image-to-image (I2I) translation is to transfer an image from a source domain to a target domain.
One major branch of this research is to formulate I2I translation based on Generative Adversarial Network (GAN)
arXiv Detail & Related papers (2023-03-07T02:29:36Z) - Unsupervised Image-to-Image Translation with Generative Prior [103.54337984566877]
Unsupervised image-to-image translation aims to learn the translation between two visual domains without paired data.
We present a novel framework, Generative Prior-guided UN Image-to-image Translation (GP-UNIT), to improve the overall quality and applicability of the translation algorithm.
arXiv Detail & Related papers (2022-04-07T17:59:23Z) - Leveraging in-domain supervision for unsupervised image-to-image
translation tasks via multi-stream generators [4.726777092009554]
We introduce two techniques to incorporate this invaluable in-domain prior knowledge for the benefit of translation quality.
We propose splitting the input data according to semantic masks, explicitly guiding the network to different behavior for the different regions of the image.
In addition, we propose training a semantic segmentation network along with the translation task, and to leverage this output as a loss term that improves robustness.
arXiv Detail & Related papers (2021-12-30T15:29:36Z) - Independent Encoder for Deep Hierarchical Unsupervised Image-to-Image
Translation [2.4826445086983475]
Main challenges of image-to-image (I2I) translation are to make the translated image realistic and retain as much information from the source domain as possible.
We propose a novel architecture, termed as IEGAN, which removes the encoder of each network and introduces an encoder that is independent of other networks.
arXiv Detail & Related papers (2021-07-06T09:18:59Z) - TransferI2I: Transfer Learning for Image-to-Image Translation from Small
Datasets [35.84311497205075]
Image-to-image (I2I) translation has matured in recent years and is able to generate high-quality realistic images.
Existing methods use transfer learning for I2I translation, but they still require the learning of millions of parameters from scratch.
We propose a new transfer learning for I2I translation (TransferI2I)
arXiv Detail & Related papers (2021-05-13T12:19:36Z) - Unpaired Image-to-Image Translation via Latent Energy Transport [61.62293304236371]
Image-to-image translation aims to preserve source contents while translating to discriminative target styles between two visual domains.
In this paper, we propose to deploy an energy-based model (EBM) in the latent space of a pretrained autoencoder for this task.
Our model is the first to be applicable to 1024$times$1024-resolution unpaired image translation.
arXiv Detail & Related papers (2020-12-01T17:18:58Z) - Unsupervised Image-to-Image Translation via Pre-trained StyleGAN2
Network [73.5062435623908]
We propose a new I2I translation method that generates a new model in the target domain via a series of model transformations.
By feeding the latent vector into the generated model, we can perform I2I translation between the source domain and target domain.
arXiv Detail & Related papers (2020-10-12T13:51:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.