Delayed takedown of illegal content on social media makes moderation ineffective
- URL: http://arxiv.org/abs/2502.08841v1
- Date: Wed, 12 Feb 2025 23:16:39 GMT
- Title: Delayed takedown of illegal content on social media makes moderation ineffective
- Authors: Bao Tran Truong, Sangyeon Kim, Gianluca Nogara, Enrico Verdolotti, Erfan Samieyan Sahneh, Florian Saurwein, Natascha Just, Luca Luceri, Silvia Giordano, Filippo Menczer,
- Abstract summary: This study models the relationship between the timeliness of illegal content removal and its prevalence, reach, and exposure on social media.
By simulating illegal content diffusion using empirical data from the DSA Transparency Database, we demonstrate that rapid takedown (within hours) significantly reduces illegal content prevalence and exposure.
While these findings support tight takedown deadlines for content removal, such deadlines cannot address the delay in identifying the illegal content and can adversely affect the quality of content moderation.
- Score: 4.4134057281132195
- License:
- Abstract: Social media platforms face legal and regulatory demands to swiftly remove illegal content, sometimes under strict takedown deadlines. However, the effects of moderation speed and the impact of takedown deadlines remain underexplored. This study models the relationship between the timeliness of illegal content removal and its prevalence, reach, and exposure on social media. By simulating illegal content diffusion using empirical data from the DSA Transparency Database, we demonstrate that rapid takedown (within hours) significantly reduces illegal content prevalence and exposure, while longer delays decrease the effectiveness of moderation efforts. While these findings support tight takedown deadlines for content removal, such deadlines cannot address the delay in identifying the illegal content and can adversely affect the quality of content moderation.
Related papers
- Advancing Content Moderation: Evaluating Large Language Models for Detecting Sensitive Content Across Text, Images, and Videos [0.1399948157377307]
Governments, educators, and parents are often at odds with media platforms about how to regulate, control, and limit the spread of such content.
Techniques from natural language processing and computer vision have been used widely to automatically identify and filter out sensitive content.
More sophisticated algorithms for understanding the context of both text and image may open rooms for improvement in content censorship.
arXiv Detail & Related papers (2024-11-26T05:29:18Z) - Explainability and Hate Speech: Structured Explanations Make Social Media Moderators Faster [72.84926097773578]
We investigate the effect of explanations on the speed of real-world moderators.
Our experiments show that while generic explanations do not affect their speed and are often ignored, structured explanations lower moderators' decision making time by 7.4%.
arXiv Detail & Related papers (2024-06-06T14:23:10Z) - Algorithmic Arbitrariness in Content Moderation [1.4849645397321183]
We show how content moderation tools can arbitrarily classify samples as toxic.
We discuss these findings in terms of human rights set out by the International Covenant on Civil and Political Rights (ICCPR)
Our study underscores the need to identify and increase the transparency of arbitrariness in content moderation applications.
arXiv Detail & Related papers (2024-02-26T19:27:00Z) - Content Moderation on Social Media in the EU: Insights From the DSA
Transparency Database [0.0]
Digital Services Act (DSA) requires large social media platforms in the EU to provide clear and specific information whenever they restrict access to certain content.
Statements of Reasons (SoRs) are collected in the DSA Transparency Database to ensure transparency and scrutiny of content moderation decisions.
We empirically analyze 156 million SoRs within an observation period of two months to provide an early look at content moderation decisions of social media platforms in the EU.
arXiv Detail & Related papers (2023-12-07T16:56:19Z) - Why Should This Article Be Deleted? Transparent Stance Detection in
Multilingual Wikipedia Editor Discussions [47.944081120226905]
We construct a novel dataset of Wikipedia editor discussions along with their reasoning in three languages.
The dataset contains the stances of the editors (keep, delete, merge, comment), along with the stated reason, and a content moderation policy, for each edit decision.
We demonstrate that stance and corresponding reason (policy) can be predicted jointly with a high degree of accuracy, adding transparency to the decision-making process.
arXiv Detail & Related papers (2023-10-09T15:11:02Z) - An Image is Worth a Thousand Toxic Words: A Metamorphic Testing
Framework for Content Moderation Software [64.367830425115]
Social media platforms are being increasingly misused to spread toxic content, including hate speech, malicious advertising, and pornography.
Despite tremendous efforts in developing and deploying content moderation methods, malicious users can evade moderation by embedding texts into images.
We propose a metamorphic testing framework for content moderation software.
arXiv Detail & Related papers (2023-08-18T20:33:06Z) - Degeneration-Tuning: Using Scrambled Grid shield Unwanted Concepts from
Stable Diffusion [106.42918868850249]
We propose a novel strategy named textbfDegeneration-Tuning (DT) to shield contents of unwanted concepts from SD weights.
As this adaptation occurs at the level of the model's weights, the SD, after DT, can be grafted onto other conditional diffusion frameworks like ControlNet to shield unwanted concepts.
arXiv Detail & Related papers (2023-08-02T03:34:44Z) - The Impact of Disinformation on a Controversial Debate on Social Media [1.299941371793082]
We study how pervasive is the presence of disinformation in the Italian debate around immigration on Twitter.
By characterising the Twitter users with an textitUntrustworthiness score, we are able to see that such bad information consumption habits are not equally distributed across the users.
arXiv Detail & Related papers (2021-06-30T10:29:07Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - Quantifying the Vulnerabilities of the Online Public Square to Adversarial Manipulation Tactics [43.98568073610101]
We use a social media model to quantify the impacts of several adversarial manipulation tactics on the quality of content.
We find that the presence of influential accounts, a hallmark of social media, exacerbates the vulnerabilities of online communities to manipulation.
These insights suggest countermeasures that platforms could employ to increase the resilience of social media users to manipulation.
arXiv Detail & Related papers (2019-07-13T21:12:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.