ISF-GAN: An Implicit Style Function for High-Resolution Image-to-Image
Translation
- URL: http://arxiv.org/abs/2109.12492v1
- Date: Sun, 26 Sep 2021 04:51:39 GMT
- Title: ISF-GAN: An Implicit Style Function for High-Resolution Image-to-Image
Translation
- Authors: Yahui Liu, Yajing Chen, Linchao Bao, Nicu Sebe, Bruno Lepri, Marco De
Nadai
- Abstract summary: This work proposes an implicit style function (ISF) to straightforwardly achieve multi-modal and multi-domain image-to-image translation.
Our results in human face and animal manipulations show significantly improved results over the baselines.
Our model enables cost-effective multi-modal unsupervised image-to-image translations at high resolution using pre-trained unconditional GANs.
- Score: 55.47515538020578
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, there has been an increasing interest in image editing methods that
employ pre-trained unconditional image generators (e.g., StyleGAN). However,
applying these methods to translate images to multiple visual domains remains
challenging. Existing works do not often preserve the domain-invariant part of
the image (e.g., the identity in human face translations), they do not usually
handle multiple domains, or do not allow for multi-modal translations. This
work proposes an implicit style function (ISF) to straightforwardly achieve
multi-modal and multi-domain image-to-image translation from pre-trained
unconditional generators. The ISF manipulates the semantics of an input latent
code to make the image generated from it lying in the desired visual domain.
Our results in human face and animal manipulations show significantly improved
results over the baselines. Our model enables cost-effective multi-modal
unsupervised image-to-image translations at high resolution using pre-trained
unconditional GANs. The code and data are available at:
\url{https://github.com/yhlleo/stylegan-mmuit}.
Related papers
- Unsupervised Image-to-Image Translation with Generative Prior [103.54337984566877]
Unsupervised image-to-image translation aims to learn the translation between two visual domains without paired data.
We present a novel framework, Generative Prior-guided UN Image-to-image Translation (GP-UNIT), to improve the overall quality and applicability of the translation algorithm.
arXiv Detail & Related papers (2022-04-07T17:59:23Z) - Multi-domain Unsupervised Image-to-Image Translation with Appearance
Adaptive Convolution [62.4972011636884]
We propose a novel multi-domain unsupervised image-to-image translation (MDUIT) framework.
We exploit the decomposed content feature and appearance adaptive convolution to translate an image into a target appearance.
We show that the proposed method produces visually diverse and plausible results in multiple domains compared to the state-of-the-art methods.
arXiv Detail & Related papers (2022-02-06T14:12:34Z) - Separating Content and Style for Unsupervised Image-to-Image Translation [20.44733685446886]
Unsupervised image-to-image translation aims to learn the mapping between two visual domains with unpaired samples.
We propose to separate the content code and style code simultaneously in a unified framework.
Based on the correlation between the latent features and the high-level domain-invariant tasks, the proposed framework demonstrates superior performance.
arXiv Detail & Related papers (2021-10-27T12:56:50Z) - Multiple GAN Inversion for Exemplar-based Image-to-Image Translation [0.0]
We propose Multiple GAN Inversion for Exemplar-based Image-to-Image Translation.
Our novel Multiple GAN Inversion avoids human intervention using a self-deciding algorithm in choosing the number of layers.
Experimental results shows the advantage of the proposed method compared to existing state-of-the-art exemplar-based image-to-image translation methods.
arXiv Detail & Related papers (2021-03-26T13:46:14Z) - Unpaired Image-to-Image Translation via Latent Energy Transport [61.62293304236371]
Image-to-image translation aims to preserve source contents while translating to discriminative target styles between two visual domains.
In this paper, we propose to deploy an energy-based model (EBM) in the latent space of a pretrained autoencoder for this task.
Our model is the first to be applicable to 1024$times$1024-resolution unpaired image translation.
arXiv Detail & Related papers (2020-12-01T17:18:58Z) - Retrieval Guided Unsupervised Multi-domain Image-to-Image Translation [59.73535607392732]
Image to image translation aims to learn a mapping that transforms an image from one visual domain to another.
We propose the use of an image retrieval system to assist the image-to-image translation task.
arXiv Detail & Related papers (2020-08-11T20:11:53Z) - COCO-FUNIT: Few-Shot Unsupervised Image Translation with a Content
Conditioned Style Encoder [70.23358875904891]
Unsupervised image-to-image translation aims to learn a mapping of an image in a given domain to an analogous image in a different domain.
We propose a new few-shot image translation model, COCO-FUNIT, which computes the style embedding of the example images conditioned on the input image.
Our model shows effectiveness in addressing the content loss problem.
arXiv Detail & Related papers (2020-07-15T02:01:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.