Seeding Diversity into AI Art
- URL: http://arxiv.org/abs/2205.00804v1
- Date: Mon, 2 May 2022 10:40:52 GMT
- Title: Seeding Diversity into AI Art
- Authors: Marvin Zammit, Antonios Liapis and Georgios N. Yannakakis
- Abstract summary: generative adversarial networks (GANs) that create a single image, in a vacuum, lack a concept of novelty regarding how their product differs from previously created ones.
We envision that an algorithm that combines the novelty preservation mechanisms in evolutionary algorithms with the power of GANs can deliberately guide its creative process towards output that is both good and novel.
- Score: 1.393683063795544
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper argues that generative art driven by conformance to a visual
and/or semantic corpus lacks the necessary criteria to be considered creative.
Among several issues identified in the literature, we focus on the fact that
generative adversarial networks (GANs) that create a single image, in a vacuum,
lack a concept of novelty regarding how their product differs from previously
created ones. We envision that an algorithm that combines the novelty
preservation mechanisms in evolutionary algorithms with the power of GANs can
deliberately guide its creative process towards output that is both good and
novel. In this paper, we use recent advances in image generation based on
semantic prompts using OpenAI's CLIP model, interrupting the GAN's iterative
process with short cycles of evolutionary divergent search. The results of
evolution are then used to continue the GAN's iterative process; we hypothesise
that this intervention will lead to more novel outputs. Testing our hypothesis
using novelty search with local competition, a quality-diversity evolutionary
algorithm that can increase visual diversity while maintaining quality in the
form of adherence to the semantic prompt, we explore how different notions of
visual diversity can affect both the process and the product of the algorithm.
Results show that even a simplistic measure of visual diversity can help
counter a drift towards similar images caused by the GAN. This first experiment
opens a new direction for introducing higher intentionality and a more nuanced
drive for GANs.
Related papers
- Semantic-Aligned Adversarial Evolution Triangle for High-Transferability Vision-Language Attack [51.16384207202798]
Vision-language pre-training models are vulnerable to multimodal adversarial examples (AEs)
Previous approaches augment image-text pairs to enhance diversity within the adversarial example generation process.
We propose sampling from adversarial evolution triangles composed of clean, historical, and current adversarial examples to enhance adversarial diversity.
arXiv Detail & Related papers (2024-11-04T23:07:51Z) - DreamCreature: Crafting Photorealistic Virtual Creatures from
Imagination [140.1641573781066]
We introduce a novel task, Virtual Creatures Generation: Given a set of unlabeled images of the target concepts, we aim to train a T2I model capable of creating new, hybrid concepts.
We propose a new method called DreamCreature, which identifies and extracts the underlying sub-concepts.
The T2I thus adapts to generate novel concepts with faithful structures and photorealistic appearance.
arXiv Detail & Related papers (2023-11-27T01:24:31Z) - ConceptLab: Creative Concept Generation using VLM-Guided Diffusion Prior
Constraints [56.824187892204314]
We present the task of creative text-to-image generation, where we seek to generate new members of a broad category.
We show that the creative generation problem can be formulated as an optimization process over the output space of the diffusion prior.
We incorporate a question-answering Vision-Language Model (VLM) that adaptively adds new constraints to the optimization problem, encouraging the model to discover increasingly more unique creations.
arXiv Detail & Related papers (2023-08-03T17:04:41Z) - Real-World Image Variation by Aligning Diffusion Inversion Chain [53.772004619296794]
A domain gap exists between generated images and real-world images, which poses a challenge in generating high-quality variations of real-world images.
We propose a novel inference pipeline called Real-world Image Variation by ALignment (RIVAL)
Our pipeline enhances the generation quality of image variations by aligning the image generation process to the source image's inversion chain.
arXiv Detail & Related papers (2023-05-30T04:09:47Z) - Augmenting Character Designers Creativity Using Generative Adversarial
Networks [0.0]
Generative Adversarial Networks (GANs) continue to attract the attention of researchers in different fields.
Most recent GANs are focused on realism, however, generating hyper-realistic output is not a priority for some domains.
We present a comparison between different GAN architectures and their performance when trained from scratch on a new visual characters dataset.
We also explore alternative techniques, such as transfer learning and data augmentation, to overcome computational resource limitations.
arXiv Detail & Related papers (2023-05-28T10:52:03Z) - Creative Discovery using QD Search [4.941630596191806]
This paper introduces a method that combines evolutionary optimisation with AI-based image classification to perform quality-diversity search.
We tested our method on a generative system that produces abstract drawings.
arXiv Detail & Related papers (2023-05-08T05:11:02Z) - Investigating GANsformer: A Replication Study of a State-of-the-Art
Image Generation Model [0.0]
We reproduce and evaluate a novel variation of the original GAN network, the GANformer.
Due to resources and time limitations, we had to constrain the network's training times, dataset types, and sizes.
arXiv Detail & Related papers (2023-03-15T12:51:16Z) - Evade the Trap of Mediocrity: Promoting Diversity and Novelty in Text
Generation via Concentrating Attention [85.5379146125199]
Powerful Transformer architectures have proven superior in generating high-quality sentences.
In this work, we find that sparser attention values in Transformer could improve diversity.
We introduce a novel attention regularization loss to control the sharpness of the attention distribution.
arXiv Detail & Related papers (2022-11-14T07:53:16Z) - Rethinking conditional GAN training: An approach using geometrically
structured latent manifolds [58.07468272236356]
Conditional GANs (cGAN) suffer from critical drawbacks such as the lack of diversity in generated outputs.
We propose a novel training mechanism that increases both the diversity and the visual quality of a vanilla cGAN.
arXiv Detail & Related papers (2020-11-25T22:54:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.