AI-Generated Faces in the Real World: A Large-Scale Case Study of Twitter Profile Images
- URL: http://arxiv.org/abs/2404.14244v3
- Date: Thu, 03 Oct 2024 08:22:20 GMT
- Title: AI-Generated Faces in the Real World: A Large-Scale Case Study of Twitter Profile Images
- Authors: Jonas Ricker, Dennis Assenmacher, Thorsten Holz, Asja Fischer, Erwin Quiring,
- Abstract summary: We conduct the first large-scale investigation of the prevalence of AI-generated profile pictures on Twitter.
Our analysis of nearly 15 million Twitter profile pictures shows that 0.052% were artificially generated, confirming their notable presence on the platform.
The results also reveal several motives, including spamming and political amplification campaigns.
- Score: 26.891299948581782
- License:
- Abstract: Recent advances in the field of generative artificial intelligence (AI) have blurred the lines between authentic and machine-generated content, making it almost impossible for humans to distinguish between such media. One notable consequence is the use of AI-generated images for fake profiles on social media. While several types of disinformation campaigns and similar incidents have been reported in the past, a systematic analysis has been lacking. In this work, we conduct the first large-scale investigation of the prevalence of AI-generated profile pictures on Twitter. We tackle the challenges of a real-world measurement study by carefully integrating various data sources and designing a multi-stage detection pipeline. Our analysis of nearly 15 million Twitter profile pictures shows that 0.052% were artificially generated, confirming their notable presence on the platform. We comprehensively examine the characteristics of these accounts and their tweet content, and uncover patterns of coordinated inauthentic behavior. The results also reveal several motives, including spamming and political amplification campaigns. Our research reaffirms the need for effective detection and mitigation strategies to cope with the potential negative effects of generative AI in the future.
Related papers
- Zero-Shot Detection of AI-Generated Images [54.01282123570917]
We propose a zero-shot entropy-based detector (ZED) to detect AI-generated images.
Inspired by recent works on machine-generated text detection, our idea is to measure how surprising the image under analysis is compared to a model of real images.
ZED achieves an average improvement of more than 3% over the SoTA in terms of accuracy.
arXiv Detail & Related papers (2024-09-24T08:46:13Z) - AI-Face: A Million-Scale Demographically Annotated AI-Generated Face Dataset and Fairness Benchmark [12.368133562194267]
We introduce the AI-Face dataset, the first million-scale demographically annotated AI-generated face image dataset.
Based on this dataset, we conduct the first comprehensive fairness benchmark to assess various AI face detectors.
arXiv Detail & Related papers (2024-06-02T15:51:33Z) - Characteristics and prevalence of fake social media profiles with AI-generated faces [7.444681337745949]
Recent advancements in generative artificial intelligence (AI) have raised concerns about their potential to create convincing fake social media accounts.
We present a systematic analysis of Twitter accounts using human faces generated by Generative Adrial Networks (GANs) for their profile pictures.
We show that they are used to spread scams, spam, and amplify coordinated messages, among other inauthentic activities.
arXiv Detail & Related papers (2024-01-05T04:10:46Z) - Invisible Relevance Bias: Text-Image Retrieval Models Prefer AI-Generated Images [67.18010640829682]
We show that AI-generated images introduce an invisible relevance bias to text-image retrieval models.
The inclusion of AI-generated images in the training data of the retrieval models exacerbates the invisible relevance bias.
We propose an effective training method aimed at alleviating the invisible relevance bias.
arXiv Detail & Related papers (2023-11-23T16:22:58Z) - Finding AI-Generated Faces in the Wild [9.390562437823078]
We focus on a more narrow task of distinguishing a real face from an AI-generated face.
This is particularly applicable when tackling inauthentic online accounts with a fake user profile photo.
We show that by focusing on only faces, a more resilient and general-purpose artifact can be detected.
arXiv Detail & Related papers (2023-11-14T22:46:01Z) - DeepfakeArt Challenge: A Benchmark Dataset for Generative AI Art Forgery and Data Poisoning Detection [57.51313366337142]
There has been growing concern over the use of generative AI for malicious purposes.
In the realm of visual content synthesis using generative AI, key areas of significant concern has been image forgery and data poisoning.
We introduce the DeepfakeArt Challenge, a large-scale challenge benchmark dataset designed specifically to aid in the building of machine learning algorithms for generative AI art forgery and data poisoning detection.
arXiv Detail & Related papers (2023-06-02T05:11:27Z) - Seeing is not always believing: Benchmarking Human and Model Perception
of AI-Generated Images [66.20578637253831]
There is a growing concern that the advancement of artificial intelligence (AI) technology may produce fake photos.
This study aims to comprehensively evaluate agents for distinguishing state-of-the-art AI-generated visual content.
arXiv Detail & Related papers (2023-04-25T17:51:59Z) - Open-Eye: An Open Platform to Study Human Performance on Identifying
AI-Synthesized Faces [51.56417104929796]
We develop an online platform called Open-eye to study the human performance of AI-synthesized faces detection.
We describe the design and workflow of the Open-eye in this paper.
arXiv Detail & Related papers (2022-05-13T14:30:59Z) - Twitter-COMMs: Detecting Climate, COVID, and Military Multimodal
Misinformation [83.2079454464572]
This paper describes our approach to the Image-Text Inconsistency Detection challenge of the DARPA Semantic Forensics (SemaFor) Program.
We collect Twitter-COMMs, a large-scale multimodal dataset with 884k tweets relevant to the topics of Climate Change, COVID-19, and Military Vehicles.
We train our approach, based on the state-of-the-art CLIP model, leveraging automatically generated random and hard negatives.
arXiv Detail & Related papers (2021-12-16T03:37:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.