CFFT-GAN: Cross-domain Feature Fusion Transformer for Exemplar-based
Image Translation
- URL: http://arxiv.org/abs/2302.01608v1
- Date: Fri, 3 Feb 2023 09:11:50 GMT
- Title: CFFT-GAN: Cross-domain Feature Fusion Transformer for Exemplar-based
Image Translation
- Authors: Tianxiang Ma, Bingchuan Li, Wei Liu, Miao Hua, Jing Dong, Tieniu Tan
- Abstract summary: We propose a more general learning approach by considering two domain features as a whole.
Specifically, we propose a Cross-domain Feature Fusion Transformer (CFFT) to learn inter- and intra-domain feature fusion.
Based on CFFT, the proposed CFFT-GAN works well on exemplar-based image translation.
- Score: 55.48699434634843
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Exemplar-based image translation refers to the task of generating images with
the desired style, while conditioning on certain input image. Most of the
current methods learn the correspondence between two input domains and lack the
mining of information within the domains. In this paper, we propose a more
general learning approach by considering two domain features as a whole and
learning both inter-domain correspondence and intra-domain potential
information interactions. Specifically, we propose a Cross-domain Feature
Fusion Transformer (CFFT) to learn inter- and intra-domain feature fusion.
Based on CFFT, the proposed CFFT-GAN works well on exemplar-based image
translation. Moreover, CFFT-GAN is able to decouple and fuse features from
multiple domains by cascading CFFT modules. We conduct rich quantitative and
qualitative experiments on several image translation tasks, and the results
demonstrate the superiority of our approach compared to state-of-the-art
methods. Ablation studies show the importance of our proposed CFFT. Application
experimental results reflect the potential of our method.
Related papers
- Frequency-Controlled Diffusion Model for Versatile Text-Guided Image-to-Image Translation [17.30877810859863]
Large-scale text-to-image (T2I) diffusion models have emerged as a powerful tool for image-to-image translation (I2I)
This paper proposes frequency-controlled diffusion model (FCDiffusion), an end-to-end diffusion-based framework.
arXiv Detail & Related papers (2024-07-03T11:05:19Z) - A Dual Domain Multi-exposure Image Fusion Network based on the
Spatial-Frequency Integration [57.14745782076976]
Multi-exposure image fusion aims to generate a single high-dynamic image by integrating images with different exposures.
We propose a novelty perspective on multi-exposure image fusion via the Spatial-Frequency Integration Framework, named MEF-SFI.
Our method achieves visual-appealing fusion results against state-of-the-art multi-exposure image fusion approaches.
arXiv Detail & Related papers (2023-12-17T04:45:15Z) - Guided Image-to-Image Translation by Discriminator-Generator
Communication [71.86347329356244]
The goal of Image-to-image (I2I) translation is to transfer an image from a source domain to a target domain.
One major branch of this research is to formulate I2I translation based on Generative Adversarial Network (GAN)
arXiv Detail & Related papers (2023-03-07T02:29:36Z) - Multi-domain Unsupervised Image-to-Image Translation with Appearance
Adaptive Convolution [62.4972011636884]
We propose a novel multi-domain unsupervised image-to-image translation (MDUIT) framework.
We exploit the decomposed content feature and appearance adaptive convolution to translate an image into a target appearance.
We show that the proposed method produces visually diverse and plausible results in multiple domains compared to the state-of-the-art methods.
arXiv Detail & Related papers (2022-02-06T14:12:34Z) - Unsupervised Image Fusion Method based on Feature Mutual Mapping [16.64607158983448]
We propose an unsupervised adaptive image fusion method to address the above issues.
We construct a global map to measure the connections of pixels between the input source images.
Our method achieves superior performance in both visual perception and objective evaluation.
arXiv Detail & Related papers (2022-01-25T07:50:14Z) - PIT: Position-Invariant Transform for Cross-FoV Domain Adaptation [53.428312630479816]
We observe that the Field of View (FoV) gap induces noticeable instance appearance differences between the source and target domains.
Motivated by the observations, we propose the textbfPosition-Invariant Transform (PIT) to better align images in different domains.
arXiv Detail & Related papers (2021-08-16T15:16:47Z) - Continuous and Diverse Image-to-Image Translation via Signed Attribute
Vectors [120.13149176992896]
We present an effectively signed attribute vector, which enables continuous translation on diverse mapping paths across various domains.
To enhance the visual quality of continuous translation results, we generate a trajectory between two sign-symmetrical attribute vectors.
arXiv Detail & Related papers (2020-11-02T18:59:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.