Content Moderation on Social Media in the EU: Insights From the DSA
Transparency Database
- URL: http://arxiv.org/abs/2312.04431v1
- Date: Thu, 7 Dec 2023 16:56:19 GMT
- Title: Content Moderation on Social Media in the EU: Insights From the DSA
Transparency Database
- Authors: Chiara Drolsbach, Nicolas Pr\"ollochs
- Abstract summary: Digital Services Act (DSA) requires large social media platforms in the EU to provide clear and specific information whenever they restrict access to certain content.
Statements of Reasons (SoRs) are collected in the DSA Transparency Database to ensure transparency and scrutiny of content moderation decisions.
We empirically analyze 156 million SoRs within an observation period of two months to provide an early look at content moderation decisions of social media platforms in the EU.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Digital Services Act (DSA) requires large social media platforms in the
EU to provide clear and specific information whenever they remove or restrict
access to certain content. These "Statements of Reasons" (SoRs) are collected
in the DSA Transparency Database to ensure transparency and scrutiny of content
moderation decisions of the providers of online platforms. In this work, we
empirically analyze 156 million SoRs within an observation period of two months
to provide an early look at content moderation decisions of social media
platforms in the EU. Our empirical analysis yields the following main findings:
(i) There are vast differences in the frequency of content moderation across
platforms. For instance, TikTok performs more than 350 times more content
moderation decisions per user than X/Twitter. (ii) Content moderation is most
commonly applied for text and videos, whereas images and other content formats
undergo moderation less frequently. (ii) The primary reasons for moderation
include content falling outside the platform's scope of service,
illegal/harmful speech, and pornography/sexualized content, with moderation of
misinformation being relatively uncommon. (iii) The majority of rule-breaking
content is detected and decided upon via automated means rather than manual
intervention. However, X/Twitter reports that it relies solely on non-automated
methods. (iv) There is significant variation in the content moderation actions
taken across platforms. Altogether, our study implies inconsistencies in how
social media platforms implement their obligations under the DSA -- resulting
in a fragmented outcome that the DSA is meant to avoid. Our findings have
important implications for regulators to clarify existing guidelines or lay out
more specific rules that ensure common standards on how social media providers
handle rule-breaking content on their platforms.
Related papers
- Algorithmic Arbitrariness in Content Moderation [1.4849645397321183]
We show how content moderation tools can arbitrarily classify samples as toxic.
We discuss these findings in terms of human rights set out by the International Covenant on Civil and Political Rights (ICCPR)
Our study underscores the need to identify and increase the transparency of arbitrariness in content moderation applications.
arXiv Detail & Related papers (2024-02-26T19:27:00Z) - The DSA Transparency Database: Auditing Self-reported Moderation Actions by Social Media [0.4597131601929317]
We analyze all 353.12M records submitted by the eight largest social media platforms in the EU during the first 100 days of the database.
Our findings have far-reaching implications for policymakers and scholars across diverse disciplines.
arXiv Detail & Related papers (2023-12-16T00:02:49Z) - Content Moderation and the Formation of Online Communities: A
Theoretical Framework [7.900694093691988]
We study the impact of content moderation policies in online communities.
We first characterize the effectiveness of a natural class of moderation policies for creating and sustaining stable communities.
arXiv Detail & Related papers (2023-10-16T16:49:44Z) - Why Should This Article Be Deleted? Transparent Stance Detection in
Multilingual Wikipedia Editor Discussions [47.944081120226905]
We construct a novel dataset of Wikipedia editor discussions along with their reasoning in three languages.
The dataset contains the stances of the editors (keep, delete, merge, comment), along with the stated reason, and a content moderation policy, for each edit decision.
We demonstrate that stance and corresponding reason (policy) can be predicted jointly with a high degree of accuracy, adding transparency to the decision-making process.
arXiv Detail & Related papers (2023-10-09T15:11:02Z) - An Image is Worth a Thousand Toxic Words: A Metamorphic Testing
Framework for Content Moderation Software [64.367830425115]
Social media platforms are being increasingly misused to spread toxic content, including hate speech, malicious advertising, and pornography.
Despite tremendous efforts in developing and deploying content moderation methods, malicious users can evade moderation by embedding texts into images.
We propose a metamorphic testing framework for content moderation software.
arXiv Detail & Related papers (2023-08-18T20:33:06Z) - Analyzing Norm Violations in Live-Stream Chat [49.120561596550395]
We study the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms.
We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch.
Our results show that appropriate contextual information can boost moderation performance by 35%.
arXiv Detail & Related papers (2023-05-18T05:58:27Z) - A User-Driven Framework for Regulating and Auditing Social Media [94.70018274127231]
We propose that algorithmic filtering should be regulated with respect to a flexible, user-driven baseline.
We require that the feeds a platform filters contain "similar" informational content as their respective baseline feeds.
We present an auditing procedure that checks whether a platform honors this requirement.
arXiv Detail & Related papers (2023-04-20T17:53:34Z) - SoK: Content Moderation in Social Media, from Guidelines to Enforcement,
and Research to Practice [9.356143195807064]
We study the 14 most popular social media content moderation guidelines and practices in the US.
We identify the differences between the content moderation employed in mainstream social media platforms compared to fringe platforms.
We highlight why platforms should shift from a one-size-fits-all model to a more inclusive model.
arXiv Detail & Related papers (2022-06-29T18:48:04Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - Detecting Harmful Content On Online Platforms: What Platforms Need Vs.
Where Research Efforts Go [44.774035806004214]
harmful content on online platforms comes in many different forms including hate speech, offensive language, bullying and harassment, misinformation, spam, violence, graphic content, sexual abuse, self harm, and many other.
Online platforms seek to moderate such content to limit societal harm, to comply with legislation, and to create a more inclusive environment for their users.
There is currently a dichotomy between what types of harmful content online platforms seek to curb, and what research efforts there are to automatically detect such content.
arXiv Detail & Related papers (2021-02-27T08:01:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.