Separating Content and Style for Unsupervised Image-to-Image Translation
- URL: http://arxiv.org/abs/2110.14404v1
- Date: Wed, 27 Oct 2021 12:56:50 GMT
- Title: Separating Content and Style for Unsupervised Image-to-Image Translation
- Authors: Yunfei Liu, Haofei Wang, Yang Yue, Feng Lu
- Abstract summary: Unsupervised image-to-image translation aims to learn the mapping between two visual domains with unpaired samples.
We propose to separate the content code and style code simultaneously in a unified framework.
Based on the correlation between the latent features and the high-level domain-invariant tasks, the proposed framework demonstrates superior performance.
- Score: 20.44733685446886
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised image-to-image translation aims to learn the mapping between two
visual domains with unpaired samples. Existing works focus on disentangling
domain-invariant content code and domain-specific style code individually for
multimodal purposes. However, less attention has been paid to interpreting and
manipulating the translated image. In this paper, we propose to separate the
content code and style code simultaneously in a unified framework. Based on the
correlation between the latent features and the high-level domain-invariant
tasks, the proposed framework demonstrates superior performance in multimodal
translation, interpretability and manipulation of the translated image.
Experimental results show that the proposed approach outperforms the existing
unsupervised image translation methods in terms of visual quality and
diversity.
Related papers
- Vector Quantized Image-to-Image Translation [31.65282783830092]
We propose introducing the vector quantization technique into the image-to-image translation framework.
Our framework achieves comparable performance to the state-of-the-art image-to-image translation and image extension methods.
arXiv Detail & Related papers (2022-07-27T04:22:29Z) - Unsupervised Image-to-Image Translation with Generative Prior [103.54337984566877]
Unsupervised image-to-image translation aims to learn the translation between two visual domains without paired data.
We present a novel framework, Generative Prior-guided UN Image-to-image Translation (GP-UNIT), to improve the overall quality and applicability of the translation algorithm.
arXiv Detail & Related papers (2022-04-07T17:59:23Z) - Marginal Contrastive Correspondence for Guided Image Generation [58.0605433671196]
Exemplar-based image translation establishes dense correspondences between a conditional input and an exemplar from two different domains.
Existing work builds the cross-domain correspondences implicitly by minimizing feature-wise distances across the two domains.
We design a Marginal Contrastive Learning Network (MCL-Net) that explores contrastive learning to learn domain-invariant features for realistic exemplar-based image translation.
arXiv Detail & Related papers (2022-04-01T13:55:44Z) - A Style-aware Discriminator for Controllable Image Translation [10.338078700632423]
Current image-to-image translations do not control the output domain beyond the classes used during training.
We propose a style-aware discriminator that acts as a critic as well as a style to provide conditions.
Experiments on multiple datasets verify that the proposed model outperforms current state-of-the-art image-to-image translation methods.
arXiv Detail & Related papers (2022-03-29T09:13:33Z) - Multi-domain Unsupervised Image-to-Image Translation with Appearance
Adaptive Convolution [62.4972011636884]
We propose a novel multi-domain unsupervised image-to-image translation (MDUIT) framework.
We exploit the decomposed content feature and appearance adaptive convolution to translate an image into a target appearance.
We show that the proposed method produces visually diverse and plausible results in multiple domains compared to the state-of-the-art methods.
arXiv Detail & Related papers (2022-02-06T14:12:34Z) - BalaGAN: Image Translation Between Imbalanced Domains via Cross-Modal
Transfer [53.79505340315916]
We introduce BalaGAN, specifically designed to tackle the domain imbalance problem.
We leverage the latent modalities of the richer domain to turn the image-to-image translation problem into a balanced, multi-class, and conditional translation problem.
We show that BalaGAN outperforms strong baselines of both unconditioned and style-transfer-based image-to-image translation methods.
arXiv Detail & Related papers (2020-10-05T14:16:41Z) - Retrieval Guided Unsupervised Multi-domain Image-to-Image Translation [59.73535607392732]
Image to image translation aims to learn a mapping that transforms an image from one visual domain to another.
We propose the use of an image retrieval system to assist the image-to-image translation task.
arXiv Detail & Related papers (2020-08-11T20:11:53Z) - Cross-domain Correspondence Learning for Exemplar-based Image
Translation [59.35767271091425]
We present a framework for exemplar-based image translation, which synthesizes a photo-realistic image from the input in a distinct domain.
The output has the style (e.g., color, texture) in consistency with the semantically corresponding objects in the exemplar.
We show that our method is superior to state-of-the-art methods in terms of image quality significantly.
arXiv Detail & Related papers (2020-04-12T09:10:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.