Image Synthesis with Adversarial Networks: a Comprehensive Survey and
Case Studies
- URL: http://arxiv.org/abs/2012.13736v1
- Date: Sat, 26 Dec 2020 13:30:42 GMT
- Title: Image Synthesis with Adversarial Networks: a Comprehensive Survey and
Case Studies
- Authors: Pourya Shamsolmoali, Masoumeh Zareapoor, Eric Granger, Huiyu Zhou,
Ruili Wang, M. Emre Celebi and Jie Yang
- Abstract summary: Generative Adversarial Networks (GANs) have been extremely successful in various application domains such as computer vision, medicine, and natural language processing.
GANs are powerful models for learning complex distributions to synthesize semantically meaningful samples.
Given the current fast GANs development, in this survey, we provide a comprehensive review of adversarial models for image synthesis.
- Score: 41.00383742615389
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Generative Adversarial Networks (GANs) have been extremely successful in
various application domains such as computer vision, medicine, and natural
language processing. Moreover, transforming an object or person to a desired
shape become a well-studied research in the GANs. GANs are powerful models for
learning complex distributions to synthesize semantically meaningful samples.
However, there is a lack of comprehensive review in this field, especially lack
of a collection of GANs loss-variant, evaluation metrics, remedies for diverse
image generation, and stable training. Given the current fast GANs development,
in this survey, we provide a comprehensive review of adversarial models for
image synthesis. We summarize the synthetic image generation methods, and
discuss the categories including image-to-image translation, fusion image
generation, label-to-image mapping, and text-to-image translation. We organize
the literature based on their base models, developed ideas related to
architectures, constraints, loss functions, evaluation metrics, and training
datasets. We present milestones of adversarial models, review an extensive
selection of previous works in various categories, and present insights on the
development route from the model-based to data-driven methods. Further, we
highlight a range of potential future research directions. One of the unique
features of this review is that all software implementations of these GAN
methods and datasets have been collected and made available in one place at
https://github.com/pshams55/GAN-Case-Study.
Related papers
- RenAIssance: A Survey into AI Text-to-Image Generation in the Era of
Large Model [93.8067369210696]
Text-to-image generation (TTI) refers to the usage of models that could process text input and generate high fidelity images based on text descriptions.
Diffusion models are one prominent type of generative model used for the generation of images through the systematic introduction of noises with repeating steps.
In the era of large models, scaling up model size and the integration with large language models have further improved the performance of TTI models.
arXiv Detail & Related papers (2023-09-02T03:27:20Z) - BioGAN: An unpaired GAN-based image to image translation model for
microbiological images [1.6427658855248812]
We develop an unpaired GAN-based (Generative Adversarial Network) image to image translation model for microbiological images.
We propose a novel design for a GAN model, BioGAN, by utilizing Adversarial and Perceptual loss in order to transform high level features of laboratory-taken images into field images.
arXiv Detail & Related papers (2023-06-09T19:30:49Z) - Investigating GANsformer: A Replication Study of a State-of-the-Art
Image Generation Model [0.0]
We reproduce and evaluate a novel variation of the original GAN network, the GANformer.
Due to resources and time limitations, we had to constrain the network's training times, dataset types, and sizes.
arXiv Detail & Related papers (2023-03-15T12:51:16Z) - Intriguing Property and Counterfactual Explanation of GAN for Remote Sensing Image Generation [25.96740500337747]
Generative adversarial networks (GANs) have achieved remarkable progress in the natural image field.
GAN model is more sensitive to the size of training data for RS image generation than for natural image generation.
We propose two innovative adjustment schemes, namely Uniformity Regularization (UR) and Entropy Regularization (ER), to increase the information learned by the GAN model.
arXiv Detail & Related papers (2023-03-09T13:22:50Z) - Multimodal Image Synthesis and Editing: The Generative AI Era [131.9569600472503]
multimodal image synthesis and editing has become a hot research topic in recent years.
We comprehensively contextualize the advance of the recent multimodal image synthesis and editing.
We describe benchmark datasets and evaluation metrics as well as corresponding experimental results.
arXiv Detail & Related papers (2021-12-27T10:00:16Z) - Comparison and Analysis of Image-to-Image Generative Adversarial
Networks: A Survey [0.0]
Generative Adversarial Networks (GANs) have recently introduced effective methods of performing Image-to-Image translations.
In this paper, we survey and analyze eight Image-to-Image Generative Adversarial Networks: Pix2Px, CycleGAN, CoGAN, StarGAN, MUNIT, StarGAN2, DA-GAN, and Self Attention GAN.
Each of these models presented state-of-the-art results and introduced new techniques to build Image-to-Image GANs.
arXiv Detail & Related papers (2021-12-23T15:11:18Z) - InvGAN: Invertible GANs [88.58338626299837]
InvGAN, short for Invertible GAN, successfully embeds real images to the latent space of a high quality generative model.
This allows us to perform image inpainting, merging, and online data augmentation.
arXiv Detail & Related papers (2021-12-08T21:39:00Z) - Meta Internal Learning [88.68276505511922]
Internal learning for single-image generation is a framework, where a generator is trained to produce novel images based on a single image.
We propose a meta-learning approach that enables training over a collection of images, in order to model the internal statistics of the sample image more effectively.
Our results show that the models obtained are as suitable as single-image GANs for many common image applications.
arXiv Detail & Related papers (2021-10-06T16:27:38Z) - MOGAN: Morphologic-structure-aware Generative Learning from a Single
Image [59.59698650663925]
Recently proposed generative models complete training based on only one image.
We introduce a MOrphologic-structure-aware Generative Adversarial Network named MOGAN that produces random samples with diverse appearances.
Our approach focuses on internal features including the maintenance of rational structures and variation on appearance.
arXiv Detail & Related papers (2021-03-04T12:45:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.