Semantically Robust Unsupervised Image Translation for Paired Remote Sensing Images
- URL: http://arxiv.org/abs/2502.11468v1
- Date: Mon, 17 Feb 2025 05:57:57 GMT
- Title: Semantically Robust Unsupervised Image Translation for Paired Remote Sensing Images
- Authors: Sheng Fang, Kaiyu Li, Zhe Li, Jianli Zhao, Xingli Zhang,
- Abstract summary: SRUIT (Semantically Robust Unsupervised Image-to-image Translation) is proposed.
It ensures semantically robust translation and produces deterministic output.
It exploits the cross-cycle-consistent adversarial networks to translate from one to the other and recover them.
- Score: 12.466979110891078
- License:
- Abstract: Image translation for change detection or classification in bi-temporal remote sensing images is unique. Although it can acquire paired images, it is still unsupervised. Moreover, strict semantic preservation in translation is always needed instead of multimodal outputs. In response to these problems, this paper proposes a new method, SRUIT (Semantically Robust Unsupervised Image-to-image Translation), which ensures semantically robust translation and produces deterministic output. Inspired by previous works, the method explores the underlying characteristics of bi-temporal Remote Sensing images and designs the corresponding networks. Firstly, we assume that bi-temporal Remote Sensing images share the same latent space, for they are always acquired from the same land location. So SRUIT makes the generators share their high-level layers, and this constraint will compel two domain mapping to fall into the same latent space. Secondly, considering land covers of bi-temporal images could evolve into each other, SRUIT exploits the cross-cycle-consistent adversarial networks to translate from one to the other and recover them. Experimental results show that constraints of sharing weights and cross-cycle consistency enable translated images with both good perceptual image quality and semantic preservation for significant differences.
Related papers
- ITTR: Unpaired Image-to-Image Translation with Transformers [34.118637795470875]
We propose an effective and efficient architecture for unpaired Image-to-Image Translation with Transformers (ITTR)
ITTR has two main designs: 1) hybrid perception block (HPB) for token mixing from different fields receptive to utilize global semantics; 2) dual pruned self-attention (DPSA) to sharply reduce the computational complexity.
Our ITTR outperforms the state-of-the-arts for unpaired image-to-image translation on six benchmark datasets.
arXiv Detail & Related papers (2022-03-30T02:46:12Z) - Multi-domain Unsupervised Image-to-Image Translation with Appearance
Adaptive Convolution [62.4972011636884]
We propose a novel multi-domain unsupervised image-to-image translation (MDUIT) framework.
We exploit the decomposed content feature and appearance adaptive convolution to translate an image into a target appearance.
We show that the proposed method produces visually diverse and plausible results in multiple domains compared to the state-of-the-art methods.
arXiv Detail & Related papers (2022-02-06T14:12:34Z) - Smoothing the Disentangled Latent Style Space for Unsupervised
Image-to-Image Translation [56.55178339375146]
Image-to-Image (I2I) multi-domain translation models are usually evaluated also using the quality of their semantic results.
We propose a new training protocol based on three specific losses which help a translation network to learn a smooth and disentangled latent style space.
arXiv Detail & Related papers (2021-06-16T17:58:21Z) - The Spatially-Correlative Loss for Various Image Translation Tasks [69.62228639870114]
We propose a novel spatially-correlative loss that is simple, efficient and yet effective for preserving scene structure consistency.
Previous methods attempt this by using pixel-level cycle-consistency or feature-level matching losses.
We show distinct improvement over baseline models in all three modes of unpaired I2I translation: single-modal, multi-modal, and even single-image translation.
arXiv Detail & Related papers (2021-04-02T02:13:30Z) - Unpaired Image-to-Image Translation via Latent Energy Transport [61.62293304236371]
Image-to-image translation aims to preserve source contents while translating to discriminative target styles between two visual domains.
In this paper, we propose to deploy an energy-based model (EBM) in the latent space of a pretrained autoencoder for this task.
Our model is the first to be applicable to 1024$times$1024-resolution unpaired image translation.
arXiv Detail & Related papers (2020-12-01T17:18:58Z) - Old Photo Restoration via Deep Latent Space Translation [46.73615925108932]
We propose to restore old photos that suffer from severe degradation through a deep learning approach.
The degradation in real photos is complex and the domain gap between synthetic images and real old photos makes the network fail to generalize.
Specifically, we train two variational autoencoders (VAEs) to respectively transform old photos and clean photos into two latent spaces.
And the translation between these two latent spaces is learned with synthetic paired data.
arXiv Detail & Related papers (2020-09-14T08:51:53Z) - Contrastive Learning for Unpaired Image-to-Image Translation [64.47477071705866]
In image-to-image translation, each patch in the output should reflect the content of the corresponding patch in the input, independent of domain.
We propose a framework based on contrastive learning to maximize mutual information between the two.
We demonstrate that our framework enables one-sided translation in the unpaired image-to-image translation setting, while improving quality and reducing training time.
arXiv Detail & Related papers (2020-07-30T17:59:58Z) - Bringing Old Photos Back to Life [46.73615925108932]
The degradation in real photos is complex and the domain gap between synthetic images and real old photos makes the network fail to generalize.
We propose a novel triplet domain translation network by leveraging real photos along with massive synthetic image pairs.
The proposed method outperforms state-of-the-art methods in terms of visual quality for old photos restoration.
arXiv Detail & Related papers (2020-04-20T17:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.