Contrastive Unpaired Translation using Focal Loss for Patch
Classification
- URL: http://arxiv.org/abs/2109.12431v1
- Date: Sat, 25 Sep 2021 20:22:33 GMT
- Title: Contrastive Unpaired Translation using Focal Loss for Patch
Classification
- Authors: Bernard Spiegl
- Abstract summary: Contrastive Unpaired Translation is a new method for image-to-image translation.
We show that using focal loss in place of cross-entropy loss within the PatchNCE loss can improve on the model's performance.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Image-to-image translation models transfer images from input domain to output
domain in an endeavor to retain the original content of the image. Contrastive
Unpaired Translation is one of the existing methods for solving such problems.
Significant advantage of this method, compared to competitors, is the ability
to train and perform well in cases where both input and output domains are only
a single image. Another key thing that differentiates this method from its
predecessors is the usage of image patches rather than the whole images. It
also turns out that sampling negatives (patches required to calculate the loss)
from the same image achieves better results than a scenario where the negatives
are sampled from other images in the dataset. This type of approach encourages
mapping of corresponding patches to the same location in relation to other
patches (negatives) while at the same time improves the output image quality
and significantly decreases memory usage as well as the time required to train
the model compared to CycleGAN method used as a baseline. Through a series of
experiments we show that using focal loss in place of cross-entropy loss within
the PatchNCE loss can improve on the model's performance and even surpass the
current state-of-the-art model for image-to-image translation.
Related papers
- Data Attribution for Text-to-Image Models by Unlearning Synthesized Images [71.23012718682634]
The goal of data attribution for text-to-image models is to identify the training images that most influence the generation of a new image.
We propose a new approach that efficiently identifies highly-influential images.
arXiv Detail & Related papers (2024-06-13T17:59:44Z) - What can we learn about a generated image corrupting its latent
representation? [57.1841740328509]
We investigate the hypothesis that we can predict image quality based on its latent representation in the GANs bottleneck.
We achieve this by corrupting the latent representation with noise and generating multiple outputs.
arXiv Detail & Related papers (2022-10-12T14:40:32Z) - Exploring Negatives in Contrastive Learning for Unpaired Image-to-Image
Translation [12.754320302262533]
We introduce a new negative Pruning technology for Unpaired image-to-image Translation (PUT) by sparsifying and ranking the patches.
The proposed algorithm is efficient, flexible and enables the model to learn essential information between corresponding patches stably.
arXiv Detail & Related papers (2022-04-23T08:31:18Z) - Contrastive Feature Loss for Image Prediction [55.373404869092866]
Training supervised image synthesis models requires a critic to compare two images: the ground truth to the result.
We introduce an information theory based approach to measuring similarity between two images.
We show that our formulation boosts the perceptual realism of output images when used as a drop-in replacement for the L1 loss.
arXiv Detail & Related papers (2021-11-12T20:39:52Z) - A Hierarchical Transformation-Discriminating Generative Model for Few
Shot Anomaly Detection [93.38607559281601]
We devise a hierarchical generative model that captures the multi-scale patch distribution of each training image.
The anomaly score is obtained by aggregating the patch-based votes of the correct transformation across scales and image regions.
arXiv Detail & Related papers (2021-04-29T17:49:48Z) - Dual Contrastive Learning for Unsupervised Image-to-Image Translation [16.759958400617947]
Unsupervised image-to-image translation tasks aim to find a mapping between a source domain X and a target domain Y from unpaired training data.
Contrastive learning for Unpaired image-to-image Translation yields state-of-the-art results.
We propose a novel method based on contrastive learning and a dual learning setting to infer an efficient mapping between unpaired data.
arXiv Detail & Related papers (2021-04-15T18:00:22Z) - Contrastive Learning for Unpaired Image-to-Image Translation [64.47477071705866]
In image-to-image translation, each patch in the output should reflect the content of the corresponding patch in the input, independent of domain.
We propose a framework based on contrastive learning to maximize mutual information between the two.
We demonstrate that our framework enables one-sided translation in the unpaired image-to-image translation setting, while improving quality and reducing training time.
arXiv Detail & Related papers (2020-07-30T17:59:58Z) - COCO-FUNIT: Few-Shot Unsupervised Image Translation with a Content
Conditioned Style Encoder [70.23358875904891]
Unsupervised image-to-image translation aims to learn a mapping of an image in a given domain to an analogous image in a different domain.
We propose a new few-shot image translation model, COCO-FUNIT, which computes the style embedding of the example images conditioned on the input image.
Our model shows effectiveness in addressing the content loss problem.
arXiv Detail & Related papers (2020-07-15T02:01:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.