It's a Feature, Not a Bug: Measuring Creative Fluidity in Image Generators
- URL: http://arxiv.org/abs/2406.18570v1
- Date: Mon, 3 Jun 2024 08:31:29 GMT
- Title: It's a Feature, Not a Bug: Measuring Creative Fluidity in Image Generators
- Authors: Aditi Ramaswamy, Melane Navaratnarajah, Hana Chockler,
- Abstract summary: Our paper attempts to define and empirically measure one facet of creative behavior in AI, by conducting an experiment to quantify the "fluidity of prompt interpretation", or just "fluidity"
To study fluidity, we introduce a clear definition for it, (2) create chains of auto-generated prompts and images seeded with an initial "ground-truth: image", (3) measure these chains' breakage points using preexisting visual and semantic metrics, and (4) use both statistical tests and visual explanations to study these chains and determine whether the image generators used to produce them exhibit significant fluidity.
- Score: 5.639451539396458
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: With the rise of freely available image generators, AI-generated art has become the centre of a series of heated debates, one of which concerns the concept of human creativity. Can an image generation AI exhibit ``creativity'' of the same type that artists do, and if so, how does that manifest? Our paper attempts to define and empirically measure one facet of creative behavior in AI, by conducting an experiment to quantify the "fluidity of prompt interpretation", or just "fluidity", in a series of selected popular image generators. To study fluidity, we (1) introduce a clear definition for it, (2) create chains of auto-generated prompts and images seeded with an initial "ground-truth: image, (3) measure these chains' breakage points using preexisting visual and semantic metrics, and (4) use both statistical tests and visual explanations to study these chains and determine whether the image generators used to produce them exhibit significant fluidity.
Related papers
- Detecting Generated Images by Fitting Natural Image Distributions [75.31113784234877]
We propose a novel framework that exploits geometric differences between the data manifold of natural and generated images.<n>We employ a pair of functions engineered to yield consistent outputs for natural images but divergent outputs for generated ones.<n>An image is identified as generated if a transformation along its data manifold induces a significant change in the loss value of a self-supervised model pre-trained on natural images.
arXiv Detail & Related papers (2025-11-03T07:20:38Z) - Provenance of AI-Generated Images: A Vector Similarity and Blockchain-based Approach [3.632189127068905]
We propose an embedding-based AI image detection framework to distinguish AI-generated images from real (human-created) ones.<n>Our methodology is built on the hypothesis that AI-generated images demonstrate closer embedding proximity to other AI-generated content.<n>Our results confirm that moderate to high perturbations minimally impact the embedding signatures, with perturbed images maintaining close similarity matches to their original versions.
arXiv Detail & Related papers (2025-10-15T00:49:56Z) - The Iconicity of the Generated Image [22.154465616964256]
How humans interpret and produce images is influenced by the images we have been exposed to.<n>Visual generative AI models are exposed to many training images and learn to generate new images based on this.
arXiv Detail & Related papers (2025-09-19T23:59:43Z) - Detecting Discrepancies Between AI-Generated and Natural Images Using Uncertainty [91.64626435585643]
We propose a novel approach for detecting AI-generated images by leveraging predictive uncertainty to mitigate misuse and associated risks.
The motivation arises from the fundamental assumption regarding the distributional discrepancy between natural and AI-generated images.
We propose to leverage large-scale pre-trained models to calculate the uncertainty as the score for detecting AI-generated images.
arXiv Detail & Related papers (2024-12-08T11:32:25Z) - Synthesis and Perceptual Scaling of High Resolution Naturalistic Images Using Stable Diffusion [0.43748379918040853]
We create a stimulus set of photorealistic images characterized by gradual transitions.<n>For each object scene, we generate 10 variants that are ordered along a perceptual continuum.<n>This ordering is also predictive of confusability of stimuli in a working memory experiment.
arXiv Detail & Related papers (2024-10-16T20:49:19Z) - The Brittleness of AI-Generated Image Watermarking Techniques: Examining Their Robustness Against Visual Paraphrasing Attacks [5.708967043277477]
We propose a visual paraphraser that can remove watermarks from images.
The proposed visual paraphraser operates in two steps. First, it generates a caption for the given image using KOSMOS-2.
During the denoising step of the diffusion pipeline, the system generates a visually similar image that is guided by the text caption.
Our empirical findings demonstrate that visual paraphrase attacks can effectively remove watermarks from images.
arXiv Detail & Related papers (2024-08-19T22:58:30Z) - How to Trace Latent Generative Model Generated Images without Artificial Watermark? [88.04880564539836]
Concerns have arisen regarding potential misuse related to images generated by latent generative models.
We propose a latent inversion based method called LatentTracer to trace the generated images of the inspected model.
Our experiments show that our method can distinguish the images generated by the inspected model and other images with a high accuracy and efficiency.
arXiv Detail & Related papers (2024-05-22T05:33:47Z) - Counterfactual Image Editing [54.21104691749547]
Counterfactual image editing is an important task in generative AI, which asks how an image would look if certain features were different.
We formalize the counterfactual image editing task using formal language, modeling the causal relationships between latent generative factors and images.
We develop an efficient algorithm to generate counterfactual images by leveraging neural causal models.
arXiv Detail & Related papers (2024-02-07T20:55:39Z) - Invisible Relevance Bias: Text-Image Retrieval Models Prefer AI-Generated Images [67.18010640829682]
We show that AI-generated images introduce an invisible relevance bias to text-image retrieval models.
The inclusion of AI-generated images in the training data of the retrieval models exacerbates the invisible relevance bias.
We propose an effective training method aimed at alleviating the invisible relevance bias.
arXiv Detail & Related papers (2023-11-23T16:22:58Z) - Diversity and Diffusion: Observations on Synthetic Image Distributions
with Stable Diffusion [6.491645162078057]
Text-to-image (TTI) systems have made it possible to create realistic images with simple text prompts.
In all of the experiments performed to date, classifiers trained solely with synthetic images perform poorly at inference.
We find four issues that limit the usefulness of TTI systems for this task: ambiguity, adherence to prompt, lack of diversity, and inability to represent the underlying concept.
arXiv Detail & Related papers (2023-10-31T18:05:15Z) - Open Set Synthetic Image Source Attribution [9.179652505898332]
We propose a new metric learning-based approach to identify synthetic images.
Our technique works by learning transferrable embeddings capable of discriminating between generators.
We demonstrate our approach's ability to attribute the source of synthetic images in open-set scenarios.
arXiv Detail & Related papers (2023-08-22T16:37:51Z) - Seeding Diversity into AI Art [1.393683063795544]
generative adversarial networks (GANs) that create a single image, in a vacuum, lack a concept of novelty regarding how their product differs from previously created ones.
We envision that an algorithm that combines the novelty preservation mechanisms in evolutionary algorithms with the power of GANs can deliberately guide its creative process towards output that is both good and novel.
arXiv Detail & Related papers (2022-05-02T10:40:52Z) - IR-GAN: Image Manipulation with Linguistic Instruction by Increment
Reasoning [110.7118381246156]
Increment Reasoning Generative Adversarial Network (IR-GAN) aims to reason consistency between visual increment in images and semantic increment in instructions.
First, we introduce the word-level and instruction-level instruction encoders to learn user's intention from history-correlated instructions as semantic increment.
Second, we embed the representation of semantic increment into that of source image for generating target image, where source image plays the role of referring auxiliary.
arXiv Detail & Related papers (2022-04-02T07:48:39Z) - Ensembling with Deep Generative Views [72.70801582346344]
generative models can synthesize "views" of artificial images that mimic real-world variations, such as changes in color or pose.
Here, we investigate whether such views can be applied to real images to benefit downstream analysis tasks such as image classification.
We use StyleGAN2 as the source of generative augmentations and investigate this setup on classification tasks involving facial attributes, cat faces, and cars.
arXiv Detail & Related papers (2021-04-29T17:58:35Z) - Counterfactual Generative Networks [59.080843365828756]
We propose to decompose the image generation process into independent causal mechanisms that we train without direct supervision.
By exploiting appropriate inductive biases, these mechanisms disentangle object shape, object texture, and background.
We show that the counterfactual images can improve out-of-distribution with a marginal drop in performance on the original classification task.
arXiv Detail & Related papers (2021-01-15T10:23:12Z) - Toward Quantifying Ambiguities in Artistic Images [21.152039726639426]
This paper presents an approach to measuring the perceptual ambiguity of a collection of images.
Crowdworkers are asked to describe image content, after different viewing durations.
Experiments are performed using images created with Generative Adversarial Networks.
arXiv Detail & Related papers (2020-08-21T21:40:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.