Safer Prompts: Reducing IP Risk in Visual Generative AI
        - URL: http://arxiv.org/abs/2505.03338v1
 - Date: Tue, 06 May 2025 09:10:12 GMT
 - Title: Safer Prompts: Reducing IP Risk in Visual Generative AI
 - Authors: Lena Reissinger, Yuanyuan Li, Anna-Carolina Haensch, Neeraj Sarna, 
 - Abstract summary: We evaluate the effectiveness of prompt engineering techniques in mitigating IP infringement risks in image generation.<n>Our findings show that Chain of Thought Prompting and Task Instruction Prompting significantly reduce the similarity between generated images and the training data of diffusion models.
 - Score: 5.545107154611679
 - License: http://creativecommons.org/licenses/by/4.0/
 - Abstract:   Visual Generative AI models have demonstrated remarkable capability in generating high-quality images from simple inputs like text prompts. However, because these models are trained on images from diverse sources, they risk memorizing and reproducing specific content, raising concerns about intellectual property (IP) infringement. Recent advances in prompt engineering offer a cost-effective way to enhance generative AI performance. In this paper, we evaluate the effectiveness of prompt engineering techniques in mitigating IP infringement risks in image generation. Our findings show that Chain of Thought Prompting and Task Instruction Prompting significantly reduce the similarity between generated images and the training data of diffusion models, thereby lowering the risk of IP infringement. 
 
       
      
        Related papers
        - Low Resource Reconstruction Attacks Through Benign Prompts [12.077836270816622]
We devise a new attack that requires low resources, assumes little to no access to the actual training set, and identifies, seemingly, benign prompts that lead to potentially-risky image reconstruction.<n>This highlights the risk that images might even be reconstructed by an uninformed user and unintentionally.
arXiv  Detail & Related papers  (2025-07-10T17:32:26Z) - Image Can Bring Your Memory Back: A Novel Multi-Modal Guided Attack   against Image Generation Model Unlearning [28.15997901023315]
Recall is a novel adversarial framework designed to compromise the robustness of unlearned IGMs.<n>It consistently outperforms existing baselines in terms of adversarial effectiveness, computational efficiency, and semantic fidelity with the original prompt.<n>These findings reveal critical vulnerabilities in current unlearning mechanisms and underscore the need for more robust solutions.
arXiv  Detail & Related papers  (2025-07-09T02:59:01Z) - Enhancing Privacy-Utility Trade-offs to Mitigate Memorization in   Diffusion Models [62.979954692036685]
We introduce PRSS, which refines the classifier-free guidance approach in diffusion models by integrating prompt re-anchoring and semantic prompt search.<n>Our approach consistently improves the privacy-utility trade-off, establishing a new state-of-the-art.
arXiv  Detail & Related papers  (2025-04-25T02:51:23Z) - Harnessing Frequency Spectrum Insights for Image Copyright Protection   Against Diffusion Models [26.821064889438777]
We present novel evidence that diffusion-generated images faithfully preserve the statistical properties of their training data.<n>We introduce emphCoprGuard, a robust frequency domain watermarking framework to safeguard against unauthorized image usage.
arXiv  Detail & Related papers  (2025-03-14T04:27:50Z) - On the Vulnerability of Concept Erasure in Diffusion Models [13.916443687966039]
Research on machine unlearning has developed various concept erasure methods, which aim to remove the effect of unwanted data through post-hoc training.<n>We show these erasure techniques are vulnerable, where images of supposedly erased concepts can still be generated using adversarially crafted prompts.<n>We introduce RECORD, a coordinate-descent-based algorithm that discovers prompts capable of eliciting the generation of erased content.
arXiv  Detail & Related papers  (2025-02-24T17:26:01Z) - CopyJudge: Automated Copyright Infringement Identification and   Mitigation in Text-to-Image Diffusion Models [58.58208005178676]
We propose CopyJudge, a novel automated infringement identification framework.<n>We employ an abstraction-filtration-comparison test framework to assess the likelihood of infringement.<n>We introduce a general LVLM-based mitigation strategy that automatically optimize infringing prompts.
arXiv  Detail & Related papers  (2025-02-21T08:09:07Z) - Detecting Discrepancies Between AI-Generated and Natural Images Using   Uncertainty [91.64626435585643]
We propose a novel approach for detecting AI-generated images by leveraging predictive uncertainty to mitigate misuse and associated risks.<n>The motivation arises from the fundamental assumption regarding the distributional discrepancy between natural and AI-generated images.<n>We propose to leverage large-scale pre-trained models to calculate the uncertainty as the score for detecting AI-generated images.
arXiv  Detail & Related papers  (2024-12-08T11:32:25Z) - RIGID: A Training-free and Model-Agnostic Framework for Robust   AI-Generated Image Detection [60.960988614701414]
RIGID is a training-free and model-agnostic method for robust AI-generated image detection.
 RIGID significantly outperforms existing trainingbased and training-free detectors.
arXiv  Detail & Related papers  (2024-05-30T14:49:54Z) - Anonymization Prompt Learning for Facial Privacy-Preserving   Text-to-Image Generation [56.46932751058042]
We train a learnable prompt prefix for text-to-image diffusion models, which forces the model to generate anonymized facial identities.
Experiments demonstrate the successful anonymization performance of APL, which anonymizes any specific individuals without compromising the quality of non-identity-specific image generation.
arXiv  Detail & Related papers  (2024-05-27T07:38:26Z) - Protect-Your-IP: Scalable Source-Tracing and Attribution against   Personalized Generation [19.250673262185767]
We propose a unified approach for image copyright source-tracing and attribution.
We introduce an innovative watermarking-attribution method that blends proactive and passive strategies.
We have conducted experiments using various celebrity portrait series sourced online.
arXiv  Detail & Related papers  (2024-05-26T15:14:54Z) - Adversarial Prompt Tuning for Vision-Language Models [86.5543597406173]
Adversarial Prompt Tuning (AdvPT) is a technique to enhance the adversarial robustness of image encoders in Vision-Language Models (VLMs)
We demonstrate that AdvPT improves resistance against white-box and black-box adversarial attacks and exhibits a synergistic effect when combined with existing image-processing-based defense techniques.
arXiv  Detail & Related papers  (2023-11-19T07:47:43Z) - CopyScope: Model-level Copyright Infringement Quantification in the
  Diffusion Workflow [6.6282087165087304]
Copyright infringement quantification is the primary and challenging step towards AI-generated image copyright traceability.
We propose CopyScope, a new framework to quantify the infringement of AI-generated images from the model level.
arXiv  Detail & Related papers  (2023-10-13T13:08:09Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.