The Spatially-Correlative Loss for Various Image Translation Tasks
- URL: http://arxiv.org/abs/2104.00854v1
- Date: Fri, 2 Apr 2021 02:13:30 GMT
- Title: The Spatially-Correlative Loss for Various Image Translation Tasks
- Authors: Chuanxia Zheng, Tat-Jen Cham, Jianfei Cai
- Abstract summary: We propose a novel spatially-correlative loss that is simple, efficient and yet effective for preserving scene structure consistency.
Previous methods attempt this by using pixel-level cycle-consistency or feature-level matching losses.
We show distinct improvement over baseline models in all three modes of unpaired I2I translation: single-modal, multi-modal, and even single-image translation.
- Score: 69.62228639870114
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We propose a novel spatially-correlative loss that is simple, efficient and
yet effective for preserving scene structure consistency while supporting large
appearance changes during unpaired image-to-image (I2I) translation. Previous
methods attempt this by using pixel-level cycle-consistency or feature-level
matching losses, but the domain-specific nature of these losses hinder
translation across large domain gaps. To address this, we exploit the spatial
patterns of self-similarity as a means of defining scene structure. Our
spatially-correlative loss is geared towards only capturing spatial
relationships within an image rather than domain appearance. We also introduce
a new self-supervised learning method to explicitly learn spatially-correlative
maps for each specific translation task. We show distinct improvement over
baseline models in all three modes of unpaired I2I translation: single-modal,
multi-modal, and even single-image translation. This new loss can easily be
integrated into existing network architectures and thus allows wide
applicability.
Related papers
- Smooth image-to-image translations with latent space interpolations [64.8170758294427]
Multi-domain image-to-image (I2I) translations can transform a source image according to the style of a target domain.
We show that our regularization techniques can improve the state-of-the-art I2I translations by a large margin.
arXiv Detail & Related papers (2022-10-03T11:57:30Z) - Marginal Contrastive Correspondence for Guided Image Generation [58.0605433671196]
Exemplar-based image translation establishes dense correspondences between a conditional input and an exemplar from two different domains.
Existing work builds the cross-domain correspondences implicitly by minimizing feature-wise distances across the two domains.
We design a Marginal Contrastive Learning Network (MCL-Net) that explores contrastive learning to learn domain-invariant features for realistic exemplar-based image translation.
arXiv Detail & Related papers (2022-04-01T13:55:44Z) - Smoothing the Disentangled Latent Style Space for Unsupervised
Image-to-Image Translation [56.55178339375146]
Image-to-Image (I2I) multi-domain translation models are usually evaluated also using the quality of their semantic results.
We propose a new training protocol based on three specific losses which help a translation network to learn a smooth and disentangled latent style space.
arXiv Detail & Related papers (2021-06-16T17:58:21Z) - Flow-based Deformation Guidance for Unpaired Multi-Contrast MRI
Image-to-Image Translation [7.8333615755210175]
In this paper, we introduce a novel approach to unpaired image-to-image translation based on the invertible architecture.
We utilize the temporal information between consecutive slices to provide more constraints to the optimization for transforming one domain to another in unpaired medical images.
arXiv Detail & Related papers (2020-12-03T09:10:22Z) - Contrastive Learning for Unpaired Image-to-Image Translation [64.47477071705866]
In image-to-image translation, each patch in the output should reflect the content of the corresponding patch in the input, independent of domain.
We propose a framework based on contrastive learning to maximize mutual information between the two.
We demonstrate that our framework enables one-sided translation in the unpaired image-to-image translation setting, while improving quality and reducing training time.
arXiv Detail & Related papers (2020-07-30T17:59:58Z) - Unpaired Image-to-Image Translation using Adversarial Consistency Loss [6.900819011690599]
We propose a novel adversarial-consistency loss for image-to-image translation.
Our method achieves state-of-the-art results on three challenging tasks: glasses removal, male-to-female translation, and selfie-to-anime translation.
arXiv Detail & Related papers (2020-03-10T17:10:38Z) - Image Fine-grained Inpainting [89.17316318927621]
We present a one-stage model that utilizes dense combinations of dilated convolutions to obtain larger and more effective receptive fields.
To better train this efficient generator, except for frequently-used VGG feature matching loss, we design a novel self-guided regression loss.
We also employ a discriminator with local and global branches to ensure local-global contents consistency.
arXiv Detail & Related papers (2020-02-07T03:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.