Image-to-Image Translation with Low Resolution Conditioning
- URL: http://arxiv.org/abs/2107.11262v1
- Date: Fri, 23 Jul 2021 14:22:12 GMT
- Title: Image-to-Image Translation with Low Resolution Conditioning
- Authors: Mohamed Abderrahmen Abid, Ihsen Hedhli, Jean-Fran\c{c}ois Lalonde,
Christian Gagne
- Abstract summary: This work aims at transferring fine details from a high resolution (HR) source image to fit a coarse, low resolution (LR) image representation of the target.
This differs from previous methods that focus on translating a given image style into a target content.
Our approach relies on training the generative model to produce HR target images that both 1) share distinctive information of the associated source image; 2) correctly match the LR target image when downscaled.
- Score: 0.28675177318965034
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most image-to-image translation methods focus on learning mappings across
domains with the assumption that images share content (e.g., pose) but have
their own domain-specific information known as style. When conditioned on a
target image, such methods aim to extract the style of the target and combine
it with the content of the source image. In this work, we consider the scenario
where the target image has a very low resolution. More specifically, our
approach aims at transferring fine details from a high resolution (HR) source
image to fit a coarse, low resolution (LR) image representation of the target.
We therefore generate HR images that share features from both HR and LR inputs.
This differs from previous methods that focus on translating a given image
style into a target content, our translation approach being able to
simultaneously imitate the style and merge the structural information of the LR
target. Our approach relies on training the generative model to produce HR
target images that both 1) share distinctive information of the associated
source image; 2) correctly match the LR target image when downscaled. We
validate our method on the CelebA-HQ and AFHQ datasets by demonstrating
improvements in terms of visual quality, diversity and coverage. Qualitative
and quantitative results show that when dealing with intra-domain image
translation, our method generates more realistic samples compared to
state-of-the-art methods such as Stargan-v2
Related papers
- SCONE-GAN: Semantic Contrastive learning-based Generative Adversarial
Network for an end-to-end image translation [18.93434486338439]
SCONE-GAN is shown to be effective for learning to generate realistic and diverse scenery images.
For more realistic and diverse image generation we introduce style reference image.
We validate the proposed algorithm for image-to-image translation and stylizing outdoor images.
arXiv Detail & Related papers (2023-11-07T10:29:16Z) - Domain Agnostic Image-to-image Translation using Low-Resolution
Conditioning [6.470760375991825]
We propose a domain-agnostic i2i method for fine-grained problems, where the domains are related.
We present a novel approach that relies on training the generative model to produce images that both share distinctive information of the associated source image.
We validate our method on the CelebA-HQ and AFHQ datasets by demonstrating improvements in terms of visual quality.
arXiv Detail & Related papers (2023-05-08T19:58:49Z) - DSI2I: Dense Style for Unpaired Image-to-Image Translation [70.93865212275412]
Unpaired exemplar-based image-to-image (UEI2I) translation aims to translate a source image to a target image domain with the style of a target image exemplar.
We propose to represent style as a dense feature map, allowing for a finer-grained transfer to the source image without requiring any external semantic information.
Our results show that the translations produced by our approach are more diverse, preserve the source content better, and are closer to the exemplars when compared to the state-of-the-art methods.
arXiv Detail & Related papers (2022-12-26T18:45:25Z) - Global and Local Alignment Networks for Unpaired Image-to-Image
Translation [170.08142745705575]
The goal of unpaired image-to-image translation is to produce an output image reflecting the target domain's style.
Due to the lack of attention to the content change in existing methods, semantic information from source images suffers from degradation during translation.
We introduce a novel approach, Global and Local Alignment Networks (GLA-Net)
Our method effectively generates sharper and more realistic images than existing approaches.
arXiv Detail & Related papers (2021-11-19T18:01:54Z) - Unpaired Image-to-Image Translation via Latent Energy Transport [61.62293304236371]
Image-to-image translation aims to preserve source contents while translating to discriminative target styles between two visual domains.
In this paper, we propose to deploy an energy-based model (EBM) in the latent space of a pretrained autoencoder for this task.
Our model is the first to be applicable to 1024$times$1024-resolution unpaired image translation.
arXiv Detail & Related papers (2020-12-01T17:18:58Z) - Unsupervised Image-to-Image Translation via Pre-trained StyleGAN2
Network [73.5062435623908]
We propose a new I2I translation method that generates a new model in the target domain via a series of model transformations.
By feeding the latent vector into the generated model, we can perform I2I translation between the source domain and target domain.
arXiv Detail & Related papers (2020-10-12T13:51:40Z) - TriGAN: Image-to-Image Translation for Multi-Source Domain Adaptation [82.52514546441247]
We propose the first approach for Multi-Source Domain Adaptation (MSDA) based on Generative Adversarial Networks.
Our method is inspired by the observation that the appearance of a given image depends on three factors: the domain, the style and the content.
We test our approach using common MSDA benchmarks, showing that it outperforms state-of-the-art methods.
arXiv Detail & Related papers (2020-04-19T05:07:22Z) - Distilling Localization for Self-Supervised Representation Learning [82.79808902674282]
Contrastive learning has revolutionized unsupervised representation learning.
Current contrastive models are ineffective at localizing the foreground object.
We propose a data-driven approach for learning in variance to backgrounds.
arXiv Detail & Related papers (2020-04-14T16:29:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.