Domain Agnostic Image-to-image Translation using Low-Resolution
Conditioning
- URL: http://arxiv.org/abs/2305.05023v2
- Date: Thu, 11 May 2023 03:15:45 GMT
- Title: Domain Agnostic Image-to-image Translation using Low-Resolution
Conditioning
- Authors: Mohamed Abid, Arman Afrasiyabi, Ihsen Hedhli, Jean-Fran\c{c}ois
Lalonde and Christian Gagn\'e
- Abstract summary: We propose a domain-agnostic i2i method for fine-grained problems, where the domains are related.
We present a novel approach that relies on training the generative model to produce images that both share distinctive information of the associated source image.
We validate our method on the CelebA-HQ and AFHQ datasets by demonstrating improvements in terms of visual quality.
- Score: 6.470760375991825
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generally, image-to-image translation (i2i) methods aim at learning mappings
across domains with the assumption that the images used for translation share
content (e.g., pose) but have their own domain-specific information (a.k.a.
style). Conditioned on a target image, such methods extract the target style
and combine it with the source image content, keeping coherence between the
domains. In our proposal, we depart from this traditional view and instead
consider the scenario where the target domain is represented by a very
low-resolution (LR) image, proposing a domain-agnostic i2i method for
fine-grained problems, where the domains are related. More specifically, our
domain-agnostic approach aims at generating an image that combines visual
features from the source image with low-frequency information (e.g. pose,
color) of the LR target image. To do so, we present a novel approach that
relies on training the generative model to produce images that both share
distinctive information of the associated source image and correctly match the
LR target image when downscaled. We validate our method on the CelebA-HQ and
AFHQ datasets by demonstrating improvements in terms of visual quality.
Qualitative and quantitative results show that when dealing with intra-domain
image translation, our method generates realistic samples compared to
state-of-the-art methods such as StarGAN v2. Ablation studies also reveal that
our method is robust to changes in color, it can be applied to
out-of-distribution images, and it allows for manual control over the final
results.
Related papers
- SCONE-GAN: Semantic Contrastive learning-based Generative Adversarial
Network for an end-to-end image translation [18.93434486338439]
SCONE-GAN is shown to be effective for learning to generate realistic and diverse scenery images.
For more realistic and diverse image generation we introduce style reference image.
We validate the proposed algorithm for image-to-image translation and stylizing outdoor images.
arXiv Detail & Related papers (2023-11-07T10:29:16Z) - Multi-cropping Contrastive Learning and Domain Consistency for
Unsupervised Image-to-Image Translation [5.562419999563734]
We propose a novel unsupervised image-to-image translation framework based on multi-cropping contrastive learning and domain consistency, called MCDUT.
In many image-to-image translation tasks, our method achieves state-of-the-art results, and the advantages of our method have been proven through comparison experiments and ablation research.
arXiv Detail & Related papers (2023-04-24T16:20:28Z) - I2F: A Unified Image-to-Feature Approach for Domain Adaptive Semantic
Segmentation [55.633859439375044]
Unsupervised domain adaptation (UDA) for semantic segmentation is a promising task freeing people from heavy annotation work.
Key idea to tackle this problem is to perform both image-level and feature-level adaptation jointly.
This paper proposes a novel UDA pipeline for semantic segmentation that unifies image-level and feature-level adaptation.
arXiv Detail & Related papers (2023-01-03T15:19:48Z) - Global and Local Alignment Networks for Unpaired Image-to-Image
Translation [170.08142745705575]
The goal of unpaired image-to-image translation is to produce an output image reflecting the target domain's style.
Due to the lack of attention to the content change in existing methods, semantic information from source images suffers from degradation during translation.
We introduce a novel approach, Global and Local Alignment Networks (GLA-Net)
Our method effectively generates sharper and more realistic images than existing approaches.
arXiv Detail & Related papers (2021-11-19T18:01:54Z) - Image-to-Image Translation with Low Resolution Conditioning [0.28675177318965034]
This work aims at transferring fine details from a high resolution (HR) source image to fit a coarse, low resolution (LR) image representation of the target.
This differs from previous methods that focus on translating a given image style into a target content.
Our approach relies on training the generative model to produce HR target images that both 1) share distinctive information of the associated source image; 2) correctly match the LR target image when downscaled.
arXiv Detail & Related papers (2021-07-23T14:22:12Z) - Evaluation of Correctness in Unsupervised Many-to-Many Image Translation [61.44666983942965]
Unsupervised many-to-many image-to-image (UMMI2I) translation methods seek to generate a plausible example from the target domain.
We propose a set of benchmarks and metrics for the evaluation of semantic correctness of UMMI2I methods.
arXiv Detail & Related papers (2021-03-29T16:13:03Z) - Learning Unsupervised Cross-domain Image-to-Image Translation Using a
Shared Discriminator [2.1377923666134118]
Unsupervised image-to-image translation is used to transform images from a source domain to generate images in a target domain without using source-target image pairs.
We propose a new method that uses a single shared discriminator between the two GANs, which improves the overall efficacy.
Our results indicate that even without adding attention mechanisms, our method performs at par with attention-based methods and generates images of comparable quality.
arXiv Detail & Related papers (2021-02-09T08:26:23Z) - Unsupervised Image-to-Image Translation via Pre-trained StyleGAN2
Network [73.5062435623908]
We propose a new I2I translation method that generates a new model in the target domain via a series of model transformations.
By feeding the latent vector into the generated model, we can perform I2I translation between the source domain and target domain.
arXiv Detail & Related papers (2020-10-12T13:51:40Z) - TriGAN: Image-to-Image Translation for Multi-Source Domain Adaptation [82.52514546441247]
We propose the first approach for Multi-Source Domain Adaptation (MSDA) based on Generative Adversarial Networks.
Our method is inspired by the observation that the appearance of a given image depends on three factors: the domain, the style and the content.
We test our approach using common MSDA benchmarks, showing that it outperforms state-of-the-art methods.
arXiv Detail & Related papers (2020-04-19T05:07:22Z) - Cross-domain Correspondence Learning for Exemplar-based Image
Translation [59.35767271091425]
We present a framework for exemplar-based image translation, which synthesizes a photo-realistic image from the input in a distinct domain.
The output has the style (e.g., color, texture) in consistency with the semantically corresponding objects in the exemplar.
We show that our method is superior to state-of-the-art methods in terms of image quality significantly.
arXiv Detail & Related papers (2020-04-12T09:10:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.