DC-Art-GAN: Stable Procedural Content Generation using DC-GANs for
Digital Art
- URL: http://arxiv.org/abs/2209.02847v1
- Date: Tue, 6 Sep 2022 23:06:46 GMT
- Title: DC-Art-GAN: Stable Procedural Content Generation using DC-GANs for
Digital Art
- Authors: Rohit Gandikota and Nik Bear Brown
- Abstract summary: We advocate the concept of using deep generative networks with adversarial training for a stable and variant art generation.
The work mainly focuses on using the Deep Convolutional Generative Adversarial Network (DC-GAN) and explores the techniques to address the common pitfalls in GAN training.
- Score: 4.9631159466100305
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Art is an artistic method of using digital technologies as a part of the
generative or creative process. With the advent of digital currency and NFTs
(Non-Fungible Token), the demand for digital art is growing aggressively. In
this manuscript, we advocate the concept of using deep generative networks with
adversarial training for a stable and variant art generation. The work mainly
focuses on using the Deep Convolutional Generative Adversarial Network (DC-GAN)
and explores the techniques to address the common pitfalls in GAN training. We
compare various architectures and designs of DC-GANs to arrive at a
recommendable design choice for a stable and realistic generation. The main
focus of the work is to generate realistic images that do not exist in reality
but are synthesised from random noise by the proposed model. We provide visual
results of generated animal face images (some pieces of evidence showing a
blend of species) along with recommendations for training, architecture and
design choices. We also show how training image preprocessing plays a massive
role in GAN training.
Related papers
- Colour and Brush Stroke Pattern Recognition in Abstract Art using Modified Deep Convolutional Generative Adversarial Networks [0.0]
This paper describes the study of a wide distribution of abstract paintings using Generative Adrial Neural Networks(GAN)
The challenge lies in developing an efficient GAN architecture that overcomes common training pitfalls.
This paper introduces a modified-DCGAN (mDCGAN) specifically designed for high-quality artwork generation.
arXiv Detail & Related papers (2024-03-27T09:35:56Z) - Harnessing Machine Learning for Discerning AI-Generated Synthetic Images [2.6227376966885476]
We employ machine learning techniques to discern between AI-generated and genuine images.
We refine and adapt advanced deep learning architectures like ResNet, VGGNet, and DenseNet.
The experimental results were significant, demonstrating that our optimized deep learning models outperform traditional methods.
arXiv Detail & Related papers (2024-01-14T20:00:37Z) - RenAIssance: A Survey into AI Text-to-Image Generation in the Era of
Large Model [93.8067369210696]
Text-to-image generation (TTI) refers to the usage of models that could process text input and generate high fidelity images based on text descriptions.
Diffusion models are one prominent type of generative model used for the generation of images through the systematic introduction of noises with repeating steps.
In the era of large models, scaling up model size and the integration with large language models have further improved the performance of TTI models.
arXiv Detail & Related papers (2023-09-02T03:27:20Z) - There Is a Digital Art History [1.0878040851637998]
We revisit Johanna Drucker's question, "Is there a digital art history?"
We focus our analysis on two main aspects that seem to suggest a coming paradigm shift towards a "digital" art history.
arXiv Detail & Related papers (2023-08-14T21:21:03Z) - Diffusion idea exploration for art generation [0.10152838128195467]
Diffusion models have recently outperformed other generative models in image generation tasks using cross modal data as guiding information.
The initial experiments for this task of novel image generation demonstrated promising qualitative results.
arXiv Detail & Related papers (2023-07-11T02:35:26Z) - StyleAvatar3D: Leveraging Image-Text Diffusion Models for High-Fidelity
3D Avatar Generation [103.88928334431786]
We present a novel method for generating high-quality, stylized 3D avatars.
We use pre-trained image-text diffusion models for data generation and a Generative Adversarial Network (GAN)-based 3D generation network for training.
Our approach demonstrates superior performance over current state-of-the-art methods in terms of visual quality and diversity of the produced avatars.
arXiv Detail & Related papers (2023-05-30T13:09:21Z) - 3DAvatarGAN: Bridging Domains for Personalized Editable Avatars [75.31960120109106]
3D-GANs synthesize geometry and texture by training on large-scale datasets with a consistent structure.
We propose an adaptation framework, where the source domain is a pre-trained 3D-GAN, while the target domain is a 2D-GAN trained on artistic datasets.
We show a deformation-based technique for modeling exaggerated geometry of artistic domains, enabling -- as a byproduct -- personalized geometric editing.
arXiv Detail & Related papers (2023-01-06T19:58:47Z) - Draw Your Art Dream: Diverse Digital Art Synthesis with Multimodal
Guided Diffusion [78.47285788155818]
Current digital art synthesis methods usually use single-modality inputs as guidance.
We propose the multimodal guided artwork diffusion (MGAD) model, which is a diffusion-based digital artwork generation approach.
arXiv Detail & Related papers (2022-09-27T13:10:25Z) - CtlGAN: Few-shot Artistic Portraits Generation with Contrastive Transfer
Learning [77.27821665339492]
CtlGAN is a new few-shot artistic portraits generation model with a novel contrastive transfer learning strategy.
We adapt a pretrained StyleGAN in the source domain to a target artistic domain with no more than 10 artistic faces.
We propose a new encoder which embeds real faces into Z+ space and proposes a dual-path training strategy to better cope with the adapted decoder.
arXiv Detail & Related papers (2022-03-16T13:28:17Z) - InvGAN: Invertible GANs [88.58338626299837]
InvGAN, short for Invertible GAN, successfully embeds real images to the latent space of a high quality generative model.
This allows us to perform image inpainting, merging, and online data augmentation.
arXiv Detail & Related papers (2021-12-08T21:39:00Z) - Blind Motion Deblurring through SinGAN Architecture [21.104218472462907]
Blind motion deblurring involves reconstructing a sharp image from an observation that is blurry.
SinGAN is a generative model that is unconditional and could be learned from a single natural image.
arXiv Detail & Related papers (2020-11-07T06:09:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.