On-Device Watermarking: A Socio-Technical Imperative For Authenticity In The Age of Generative AI
- URL: http://arxiv.org/abs/2504.13205v1
- Date: Tue, 15 Apr 2025 20:36:52 GMT
- Title: On-Device Watermarking: A Socio-Technical Imperative For Authenticity In The Age of Generative AI
- Authors: Houssam Kherraz,
- Abstract summary: We argue that we are adopting the wrong approach, and should instead focus on watermarking via cryptographic signatures.<n>For audio-visual content, in particular, all real content is grounded in the physical world and captured via hardware sensors.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As generative AI models produce increasingly realistic output, both academia and industry are focusing on the ability to detect whether an output was generated by an AI model or not. Many of the research efforts and policy discourse are centered around robust watermarking of AI outputs. While plenty of progress has been made, all watermarking and AI detection techniques face severe limitations. In this position paper, we argue that we are adopting the wrong approach, and should instead focus on watermarking via cryptographic signatures trustworthy content rather than AI generated ones. For audio-visual content, in particular, all real content is grounded in the physical world and captured via hardware sensors. This presents a unique opportunity to watermark at the hardware layer, and we lay out a socio-technical framework and draw parallels with HTTPS certification and Blu-Ray verification protocols. While acknowledging implementation challenges, we contend that hardware-based authentication offers a more tractable path forward, particularly from a policy perspective. As generative models approach perceptual indistinguishability, the research community should be wary of being overly optimistic with AI watermarking, and we argue that AI watermarking research efforts are better spent in the text and LLM space, which are ultimately not traceable to a physical sensor.
Related papers
- Could AI Trace and Explain the Origins of AI-Generated Images and Text? [53.11173194293537]
AI-generated content is increasingly prevalent in the real world.
adversaries might exploit large multimodal models to create images that violate ethical or legal standards.
Paper reviewers may misuse large language models to generate reviews without genuine intellectual effort.
arXiv Detail & Related papers (2025-04-05T20:51:54Z) - Almost AI, Almost Human: The Challenge of Detecting AI-Polished Writing [55.2480439325792]
Misclassification can lead to false plagiarism accusations and misleading claims about AI prevalence in online content.
We systematically evaluate eleven state-of-the-art AI-text detectors using our AI-Polished-Text Evaluation dataset.
Our findings reveal that detectors frequently misclassify even minimally polished text as AI-generated, struggle to differentiate between degrees of AI involvement, and exhibit biases against older and smaller models.
arXiv Detail & Related papers (2025-02-21T18:45:37Z) - Watermarking across Modalities for Content Tracing and Generative AI [2.456311843339488]
This thesis includes the development of new watermarking techniques for images, audio, and text.
We first introduce methods for active moderation of images on social platforms.
We then develop specific techniques for AI-generated content.
arXiv Detail & Related papers (2025-02-04T18:49:50Z) - SoK: Watermarking for AI-Generated Content [112.9218881276487]
Watermarking schemes embed hidden signals within AI-generated content to enable reliable detection.<n>Watermarks can play a crucial role in enhancing AI safety and trustworthiness by combating misinformation and deception.<n>This work aims to guide researchers in advancing watermarking methods and applications, and support policymakers in addressing the broader implications of GenAI.
arXiv Detail & Related papers (2024-11-27T16:22:33Z) - SoK: On the Role and Future of AIGC Watermarking in the Era of Gen-AI [24.187726079290357]
AIGC watermarks offer an effective solution to mitigate malicious activities.
We provide a taxonomy based on the core properties of the watermark.
We discuss the functionality and security threats of AIGC watermarking.
arXiv Detail & Related papers (2024-11-18T11:26:42Z) - Certifiably Robust Image Watermark [57.546016845801134]
Generative AI raises many societal concerns such as boosting disinformation and propaganda campaigns.
Watermarking AI-generated content is a key technology to address these concerns.
We propose the first image watermarks with certified robustness guarantees against removal and forgery attacks.
arXiv Detail & Related papers (2024-07-04T17:56:04Z) - ACW: Enhancing Traceability of AI-Generated Codes Based on Watermarking [5.782554045290121]
It is desirable to know whether a piece of code is generated by AI, and which AI is the creator.
Watermarking is broadly considered a promising solution and is successfully applied for identifying AI-generated text.
We propose ACW (AI Code Watermarking), a novel method for watermarking AI-generated codes.
arXiv Detail & Related papers (2024-02-12T09:40:18Z) - Evading Watermark based Detection of AI-Generated Content [45.47476727209842]
A generative AI model can generate extremely realistic-looking content.
Watermark has been leveraged to detect AI-generated content.
A content is detected as AI-generated if a similar watermark can be decoded from it.
arXiv Detail & Related papers (2023-05-05T19:20:29Z) - Can AI-Generated Text be Reliably Detected? [50.95804851595018]
Large Language Models (LLMs) perform impressively well in various applications.
The potential for misuse of these models in activities such as plagiarism, generating fake news, and spamming has raised concern about their responsible use.
We stress-test the robustness of these AI text detectors in the presence of an attacker.
arXiv Detail & Related papers (2023-03-17T17:53:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.