Image-to-image Transformation with Auxiliary Condition
- URL: http://arxiv.org/abs/2106.13696v1
- Date: Fri, 25 Jun 2021 15:33:11 GMT
- Title: Image-to-image Transformation with Auxiliary Condition
- Authors: Robert Leer, Hessi Roma, James Amelia
- Abstract summary: We propose to introduce the label information of subjects, e.g., pose and type of objects in the training of CycleGAN, and lead it to obtain label-wise transforamtion models.
We evaluate our proposed method called Label-CycleGAN, through experiments on the digit image transformation from SVHN to MNIST and the surveillance camera image transformation from simulated to real images.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The performance of image recognition like human pose detection, trained with
simulated images would usually get worse due to the divergence between real and
simulated data. To make the distribution of a simulated image close to that of
real one, there are several works applying GAN-based image-to-image
transformation methods, e.g., SimGAN and CycleGAN. However, these methods would
not be sensitive enough to the various change in pose and shape of subjects,
especially when the training data are imbalanced, e.g., some particular poses
and shapes are minor in the training data. To overcome this problem, we propose
to introduce the label information of subjects, e.g., pose and type of objects
in the training of CycleGAN, and lead it to obtain label-wise transforamtion
models. We evaluate our proposed method called Label-CycleGAN, through
experiments on the digit image transformation from SVHN to MNIST and the
surveillance camera image transformation from simulated to real images.
Related papers
- Learning Transformations To Reduce the Geometric Shift in Object
Detection [60.20931827772482]
We tackle geometric shifts emerging from variations in the image capture process.
We introduce a self-training approach that learns a set of geometric transformations to minimize these shifts.
We evaluate our method on two different shifts, i.e., a camera's field of view (FoV) change and a viewpoint change.
arXiv Detail & Related papers (2023-01-13T11:55:30Z) - DiffuseMorph: Unsupervised Deformable Image Registration Along
Continuous Trajectory Using Diffusion Models [31.826844124173984]
We present a novel approach of diffusion model-based probabilistic image registration, called DiffuseMorph.
Our model learns the score function of the deformation between moving and fixed images.
Our method can provide flexible and accurate deformation with a capability of topology preservation.
arXiv Detail & Related papers (2021-12-09T08:41:23Z) - Extracting Deformation-Aware Local Features by Learning to Deform [3.364554138758565]
We present a new approach to compute features from still images that are robust to non-rigid deformations.
We train the model architecture end-to-end by applying non-rigid deformations to objects in a simulated environment.
Experiments show that our method outperforms state-of-the-art handcrafted, learning-based image, and RGB-D descriptors in different datasets.
arXiv Detail & Related papers (2021-11-20T15:46:33Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z) - A Hierarchical Transformation-Discriminating Generative Model for Few
Shot Anomaly Detection [93.38607559281601]
We devise a hierarchical generative model that captures the multi-scale patch distribution of each training image.
The anomaly score is obtained by aggregating the patch-based votes of the correct transformation across scales and image regions.
arXiv Detail & Related papers (2021-04-29T17:49:48Z) - Blind Motion Deblurring through SinGAN Architecture [21.104218472462907]
Blind motion deblurring involves reconstructing a sharp image from an observation that is blurry.
SinGAN is a generative model that is unconditional and could be learned from a single natural image.
arXiv Detail & Related papers (2020-11-07T06:09:16Z) - Self-Supervised Linear Motion Deblurring [112.75317069916579]
Deep convolutional neural networks are state-of-the-art for image deblurring.
We present a differentiable reblur model for self-supervised motion deblurring.
Our experiments demonstrate that self-supervised single image deblurring is really feasible.
arXiv Detail & Related papers (2020-02-10T20:15:21Z) - Learning Numerical Observers using Unsupervised Domain Adaptation [13.548174682737756]
Medical imaging systems are commonly assessed by use of objective image quality measures.
Supervised deep learning methods have been investigated to implement numerical observers for task-based image quality assessment.
labeling large amounts of experimental data to train deep neural networks is tedious, expensive, and prone to subjective errors.
arXiv Detail & Related papers (2020-02-03T22:58:28Z) - Deformation-aware Unpaired Image Translation for Pose Estimation on
Laboratory Animals [56.65062746564091]
We aim to capture the pose of neuroscience model organisms, without using any manual supervision, to study how neural circuits orchestrate behaviour.
Our key contribution is the explicit and independent modeling of appearance, shape and poses in an unpaired image translation framework.
We demonstrate improved pose estimation accuracy on Drosophila melanogaster (fruit fly), Caenorhabditis elegans (worm) and Danio rerio (zebrafish)
arXiv Detail & Related papers (2020-01-23T15:34:11Z) - Fine-grained Image-to-Image Transformation towards Visual Recognition [102.51124181873101]
We aim at transforming an image with a fine-grained category to synthesize new images that preserve the identity of the input image.
We adopt a model based on generative adversarial networks to disentangle the identity related and unrelated factors of an image.
Experiments on the CompCars and Multi-PIE datasets demonstrate that our model preserves the identity of the generated images much better than the state-of-the-art image-to-image transformation models.
arXiv Detail & Related papers (2020-01-12T05:26:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.