PREGAN: Pose Randomization and Estimation for Weakly Paired Image Style
Translation
- URL: http://arxiv.org/abs/2011.00301v2
- Date: Sun, 17 Jan 2021 07:18:56 GMT
- Title: PREGAN: Pose Randomization and Estimation for Weakly Paired Image Style
Translation
- Authors: Zexi Chen, Jiaxin Guo, Xuecheng Xu, Yunkai Wang, Yue Wang, Rong Xiong
- Abstract summary: We propose a weakly-paired setting for the style translation, where the content in the two images is aligned with errors in poses.
PREGAN is validated on both simulated and real-world collected data to show the effectiveness.
- Score: 11.623477199795037
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Utilizing the trained model under different conditions without data
annotation is attractive for robot applications. Towards this goal, one class
of methods is to translate the image style from another environment to the one
on which models are trained. In this paper, we propose a weakly-paired setting
for the style translation, where the content in the two images is aligned with
errors in poses. These images could be acquired by different sensors in
different conditions that share an overlapping region, e.g. with LiDAR or
stereo cameras, from sunny days or foggy nights. We consider this setting to be
more practical with: (i) easier labeling than the paired data; (ii) better
interpretability and detail retrieval than the unpaired data. To translate
across such images, we propose PREGAN to train a style translator by
intentionally transforming the two images with a random pose, and to estimate
the given random pose by differentiable non-trainable pose estimator given that
the more aligned in style, the better the estimated result is. Such adversarial
training enforces the network to learn the style translation, avoiding being
entangled with other variations. Finally, PREGAN is validated on both simulated
and real-world collected data to show the effectiveness. Results on down-stream
tasks, classification, road segmentation, object detection, and feature
matching show its potential for real applications.
https://github.com/wrld/PRoGAN
Related papers
- Lost in Translation: Modern Neural Networks Still Struggle With Small Realistic Image Transformations [8.248839892711478]
Deep neural networks that achieve remarkable performance in image classification can be easily fooled by tiny transformations.
We show that these approaches still fall short in robustly handling 'natural' image translations that simulate a subtle change in camera orientation.
We present Robust Inference by Crop Selection: a simple method that can be proven to achieve any desired level of consistency.
arXiv Detail & Related papers (2024-04-10T16:39:50Z) - DEADiff: An Efficient Stylization Diffusion Model with Disentangled
Representations [64.43387739794531]
Current encoder-based approaches significantly impair the text controllability of text-to-image models while transferring styles.
We introduce DEADiff to address this issue using the following two strategies.
DEAiff attains the best visual stylization results and optimal balance between the text controllability inherent in the text-to-image model and style similarity to the reference image.
arXiv Detail & Related papers (2024-03-11T17:35:23Z) - SCONE-GAN: Semantic Contrastive learning-based Generative Adversarial
Network for an end-to-end image translation [18.93434486338439]
SCONE-GAN is shown to be effective for learning to generate realistic and diverse scenery images.
For more realistic and diverse image generation we introduce style reference image.
We validate the proposed algorithm for image-to-image translation and stylizing outdoor images.
arXiv Detail & Related papers (2023-11-07T10:29:16Z) - Image-to-Image Translation for Autonomous Driving from Coarsely-Aligned
Image Pairs [57.33431586417377]
A self-driving car must be able to handle adverse weather conditions to operate safely.
In this paper, we investigate the idea of turning sensor inputs captured in an adverse condition into a benign one.
We show that our coarsely-aligned training scheme leads to a better image translation quality and improved downstream tasks.
arXiv Detail & Related papers (2022-09-23T16:03:18Z) - Unpaired Image-to-Image Translation via Latent Energy Transport [61.62293304236371]
Image-to-image translation aims to preserve source contents while translating to discriminative target styles between two visual domains.
In this paper, we propose to deploy an energy-based model (EBM) in the latent space of a pretrained autoencoder for this task.
Our model is the first to be applicable to 1024$times$1024-resolution unpaired image translation.
arXiv Detail & Related papers (2020-12-01T17:18:58Z) - An Empirical Study of the Collapsing Problem in Semi-Supervised 2D Human
Pose Estimation [80.02124918255059]
Semi-supervised learning aims to boost the accuracy of a model by exploring unlabeled images.
We learn two networks to mutually teach each other.
The more reliable predictions on easy images in each network are used to teach the other network to learn about the corresponding hard images.
arXiv Detail & Related papers (2020-11-25T03:29:52Z) - Unsupervised Image-to-Image Translation via Pre-trained StyleGAN2
Network [73.5062435623908]
We propose a new I2I translation method that generates a new model in the target domain via a series of model transformations.
By feeding the latent vector into the generated model, we can perform I2I translation between the source domain and target domain.
arXiv Detail & Related papers (2020-10-12T13:51:40Z) - Contrastive Learning for Unpaired Image-to-Image Translation [64.47477071705866]
In image-to-image translation, each patch in the output should reflect the content of the corresponding patch in the input, independent of domain.
We propose a framework based on contrastive learning to maximize mutual information between the two.
We demonstrate that our framework enables one-sided translation in the unpaired image-to-image translation setting, while improving quality and reducing training time.
arXiv Detail & Related papers (2020-07-30T17:59:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.