Hierarchy Flow For High-Fidelity Image-to-Image Translation
- URL: http://arxiv.org/abs/2308.06909v1
- Date: Mon, 14 Aug 2023 03:11:17 GMT
- Title: Hierarchy Flow For High-Fidelity Image-to-Image Translation
- Authors: Weichen Fan, Jinghuan Chen, Ziwei Liu
- Abstract summary: We propose a novel flow-based model to achieve better content preservation during translation.
Our approach achieves state-of-the-art performance, with convincing advantages in both strong- and normal-fidelity tasks.
- Score: 38.87847690777645
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Image-to-image (I2I) translation comprises a wide spectrum of tasks. Here we
divide this problem into three levels: strong-fidelity translation,
normal-fidelity translation, and weak-fidelity translation, indicating the
extent to which the content of the original image is preserved. Although
existing methods achieve good performance in weak-fidelity translation, they
fail to fully preserve the content in both strong- and normal-fidelity tasks,
e.g. sim2real, style transfer and low-level vision. In this work, we propose
Hierarchy Flow, a novel flow-based model to achieve better content preservation
during translation. Specifically, 1) we first unveil the drawbacks of standard
flow-based models when applied to I2I translation. 2) Next, we propose a new
design, namely hierarchical coupling for reversible feature transformation and
multi-scale modeling, to constitute Hierarchy Flow. 3) Finally, we present a
dedicated aligned-style loss for a better trade-off between content
preservation and stylization during translation. Extensive experiments on a
wide range of I2I translation benchmarks demonstrate that our approach achieves
state-of-the-art performance, with convincing advantages in both strong- and
normal-fidelity tasks. Code and models will be at
https://github.com/WeichenFan/HierarchyFlow.
Related papers
- StyleFlow For Content-Fixed Image to Image Translation [15.441136520005578]
StyleFlow is a new I2I translation model that consists of normalizing flows and a novel Style-Aware Normalization (SAN) module.
Our model supports both image-guided translation and multi-modal synthesis.
arXiv Detail & Related papers (2022-07-05T09:40:03Z) - Unsupervised Image-to-Image Translation with Generative Prior [103.54337984566877]
Unsupervised image-to-image translation aims to learn the translation between two visual domains without paired data.
We present a novel framework, Generative Prior-guided UN Image-to-image Translation (GP-UNIT), to improve the overall quality and applicability of the translation algorithm.
arXiv Detail & Related papers (2022-04-07T17:59:23Z) - Bi-level Feature Alignment for Versatile Image Translation and
Manipulation [88.5915443957795]
Generative adversarial networks (GANs) have achieved great success in image translation and manipulation.
High-fidelity image generation with faithful style control remains a grand challenge in computer vision.
This paper presents a versatile image translation and manipulation framework that achieves accurate semantic and style guidance.
arXiv Detail & Related papers (2021-07-07T05:26:29Z) - Smoothing the Disentangled Latent Style Space for Unsupervised
Image-to-Image Translation [56.55178339375146]
Image-to-Image (I2I) multi-domain translation models are usually evaluated also using the quality of their semantic results.
We propose a new training protocol based on three specific losses which help a translation network to learn a smooth and disentangled latent style space.
arXiv Detail & Related papers (2021-06-16T17:58:21Z) - Unpaired Image-to-Image Translation via Latent Energy Transport [61.62293304236371]
Image-to-image translation aims to preserve source contents while translating to discriminative target styles between two visual domains.
In this paper, we propose to deploy an energy-based model (EBM) in the latent space of a pretrained autoencoder for this task.
Our model is the first to be applicable to 1024$times$1024-resolution unpaired image translation.
arXiv Detail & Related papers (2020-12-01T17:18:58Z) - Unsupervised Image-to-Image Translation via Pre-trained StyleGAN2
Network [73.5062435623908]
We propose a new I2I translation method that generates a new model in the target domain via a series of model transformations.
By feeding the latent vector into the generated model, we can perform I2I translation between the source domain and target domain.
arXiv Detail & Related papers (2020-10-12T13:51:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.