Regeneration Based Training-free Attribution of Fake Images Generated by
Text-to-Image Generative Models
- URL: http://arxiv.org/abs/2403.01489v1
- Date: Sun, 3 Mar 2024 11:55:49 GMT
- Title: Regeneration Based Training-free Attribution of Fake Images Generated by
Text-to-Image Generative Models
- Authors: Meiling Li, Zhenxing Qian, Xinpeng Zhang
- Abstract summary: We present a training-free method to attribute fake images generated by text-to-image models to their source models.
By calculating and ranking the similarity of the test image and the candidate images, we can determine the source of the image.
- Score: 39.33821502730661
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Text-to-image generative models have recently garnered significant attention
due to their ability to generate images based on prompt descriptions. While
these models have shown promising performance, concerns have been raised
regarding the potential misuse of the generated fake images. In response to
this, we have presented a simple yet effective training-free method to
attribute fake images generated by text-to-image models to their source models.
Given a test image to be attributed, we first inverse the textual prompt of the
image, and then put the reconstructed prompt into different candidate models to
regenerate candidate fake images. By calculating and ranking the similarity of
the test image and the candidate images, we can determine the source of the
image. This attribution allows model owners to be held accountable for any
misuse of their models. Note that our approach does not limit the number of
candidate text-to-image generative models. Comprehensive experiments reveal
that (1) Our method can effectively attribute fake images to their source
models, achieving comparable attribution performance with the state-of-the-art
method; (2) Our method has high scalability ability, which is well adapted to
real-world attribution scenarios. (3) The proposed method yields satisfactory
robustness to common attacks, such as Gaussian blurring, JPEG compression, and
Resizing. We also analyze the factors that influence the attribution
performance, and explore the boost brought by the proposed method as a plug-in
to improve the performance of existing SOTA. We hope our work can shed some
light on the solutions to addressing the source of AI-generated images, as well
as to prevent the misuse of text-to-image generative models.
Related papers
- Data Attribution for Text-to-Image Models by Unlearning Synthesized Images [71.23012718682634]
The goal of data attribution for text-to-image models is to identify the training images that most influence the generation of a new image.
We propose a new approach that efficiently identifies highly-influential images.
arXiv Detail & Related papers (2024-06-13T17:59:44Z) - Autoregressive Model Beats Diffusion: Llama for Scalable Image Generation [52.509092010267665]
We introduce LlamaGen, a new family of image generation models that apply original next-token prediction'' paradigm of large language models to visual generation domain.
It is an affirmative answer to whether vanilla autoregressive models, e.g., Llama, without inductive biases on visual signals can achieve state-of-the-art image generation performance if scaling properly.
arXiv Detail & Related papers (2024-06-10T17:59:52Z) - How to Trace Latent Generative Model Generated Images without Artificial Watermark? [88.04880564539836]
Concerns have arisen regarding potential misuse related to images generated by latent generative models.
We propose a latent inversion based method called LatentTracer to trace the generated images of the inspected model.
Our experiments show that our method can distinguish the images generated by the inspected model and other images with a high accuracy and efficiency.
arXiv Detail & Related papers (2024-05-22T05:33:47Z) - Which Model Generated This Image? A Model-Agnostic Approach for Origin Attribution [23.974575820244944]
In this work, we study the origin attribution of generated images in a practical setting.
The goal is to check if a given image is generated by the source model.
We propose OCC-CLIP, a CLIP-based framework for few-shot one-class classification.
arXiv Detail & Related papers (2024-04-03T12:54:16Z) - Active Generation for Image Classification [45.93535669217115]
We propose to address the efficiency of image generation by focusing on the specific needs and characteristics of the model.
With a central tenet of active learning, our method, named ActGen, takes a training-aware approach to image generation.
arXiv Detail & Related papers (2024-03-11T08:45:31Z) - A Single Simple Patch is All You Need for AI-generated Image Detection [19.541645669791023]
We find that generative models tend to focus on generating the patches with rich textures to make the images more realistic.
In this paper, we propose to exploit the noise pattern of a single simple patch to identify fake images.
Our method can achieve state-of-the-art performance on public benchmarks.
arXiv Detail & Related papers (2024-02-02T03:50:45Z) - Detecting Generated Images by Real Images Only [64.12501227493765]
Existing generated image detection methods detect visual artifacts in generated images or learn discriminative features from both real and generated images by massive training.
This paper approaches the generated image detection problem from a new perspective: Start from real images.
By finding the commonality of real images and mapping them to a dense subspace in feature space, the goal is that generated images, regardless of their generative model, are then projected outside the subspace.
arXiv Detail & Related papers (2023-11-02T03:09:37Z) - WOUAF: Weight Modulation for User Attribution and Fingerprinting in Text-to-Image Diffusion Models [32.29120988096214]
This paper introduces a novel approach to model fingerprinting that assigns responsibility for the generated images.
Our method modifies generative models based on each user's unique digital fingerprint, imprinting a unique identifier onto the resultant content that can be traced back to the user.
arXiv Detail & Related papers (2023-06-07T19:44:14Z) - If at First You Don't Succeed, Try, Try Again: Faithful Diffusion-based
Text-to-Image Generation by Selection [53.320946030761796]
diffusion-based text-to-image (T2I) models can lack faithfulness to the text prompt.
We show that large T2I diffusion models are more faithful than usually assumed, and can generate images faithful to even complex prompts.
We introduce a pipeline that generates candidate images for a text prompt and picks the best one according to an automatic scoring system.
arXiv Detail & Related papers (2023-05-22T17:59:41Z) - Implementing and Experimenting with Diffusion Models for Text-to-Image
Generation [0.0]
Two models, DALL-E 2 and Imagen, have demonstrated that highly photorealistic images could be generated from a simple textual description of an image.
Text-to-image models require exceptionally large amounts of computational resources to train, as well as handling huge datasets collected from the internet.
This thesis contributes by reviewing the different approaches and techniques used by these models, and then by proposing our own implementation of a text-to-image model.
arXiv Detail & Related papers (2022-09-22T12:03:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.