Towards Intersectional Moderation: An Alternative Model of Moderation
Built on Care and Power
- URL: http://arxiv.org/abs/2305.11250v1
- Date: Thu, 18 May 2023 18:27:52 GMT
- Title: Towards Intersectional Moderation: An Alternative Model of Moderation
Built on Care and Power
- Authors: Sarah A. Gilbert
- Abstract summary: I perform a collaborative ethnography with moderators of r/AskHistorians, a community that uses an alternative moderation model.
I focus on three emblematic controversies of r/AskHistorians' alternative model of moderation.
I argue that designers should support decision-making processes and policy makers should account for the impact of sociotechnical systems.
- Score: 0.4351216340655199
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Shortcomings of current models of moderation have driven policy makers,
scholars, and technologists to speculate about alternative models of content
moderation. While alternative models provide hope for the future of online
spaces, they can fail without proper scaffolding. Community moderators are
routinely confronted with similar issues and have therefore found creative ways
to navigate these challenges. Learning more about the decisions these
moderators make, the challenges they face, and where they are successful can
provide valuable insight into how to ensure alternative moderation models are
successful.
In this study, I perform a collaborative ethnography with moderators of
r/AskHistorians, a community that uses an alternative moderation model,
highlighting the importance of accounting for power in moderation. Drawing from
Black feminist theory, I call this "intersectional moderation." I focus on
three controversies emblematic of r/AskHistorians' alternative model of
moderation: a disagreement over a moderation decision; a collaboration to fight
racism on Reddit; and a period of intense turmoil and its impact on policy.
Through this evidence I show how volunteer moderators navigated multiple layers
of power through care work. To ensure the successful implementation of
intersectional moderation, I argue that designers should support
decision-making processes and policy makers should account for the impact of
the sociotechnical systems in which moderators work.
Related papers
- Venire: A Machine Learning-Guided Panel Review System for Community Content Moderation [17.673993032146527]
We develop Venire, an ML-backed system for panel review on Reddit.
Venire uses a machine learning model trained on log data to identify the cases where moderators are most likely to disagree.
We show that Venire is able to improve decision consistency and surface latent disagreements.
arXiv Detail & Related papers (2024-10-30T20:39:34Z) - Fundamental Problems With Model Editing: How Should Rational Belief Revision Work in LLMs? [61.68363765350178]
This paper critiques the standard formulation of the model editing problem and proposes a formal testbed for model editing research.
We first describe 12 open problems with model editing, based on challenges with (1) defining the problem, (2) developing benchmarks, and (3) assuming LLMs have editable beliefs in the first place.
Next, we introduce a semi-synthetic dataset for model editing based on Wikidata, where we can evaluate edits against labels given by an idealized Bayesian agent.
arXiv Detail & Related papers (2024-06-27T17:33:03Z) - Explainability and Hate Speech: Structured Explanations Make Social Media Moderators Faster [72.84926097773578]
We investigate the effect of explanations on the speed of real-world moderators.
Our experiments show that while generic explanations do not affect their speed and are often ignored, structured explanations lower moderators' decision making time by 7.4%.
arXiv Detail & Related papers (2024-06-06T14:23:10Z) - Concept Arithmetics for Circumventing Concept Inhibition in Diffusion Models [58.065255696601604]
We use compositional property of diffusion models, which allows to leverage multiple prompts in a single image generation.
We argue that it is essential to consider all possible approaches to image generation with diffusion models that can be employed by an adversary.
arXiv Detail & Related papers (2024-04-21T16:35:16Z) - Content Moderation Justice and Fairness on Social Media: Comparisons
Across Different Contexts and Platforms [23.735552021636245]
We conduct an online experiment on 200 American social media users of Reddit and Twitter.
We find that retributive moderation delivers higher justice and fairness for commercially moderated platforms in illegal violations.
We discuss the opportunities for platform policymaking to improve moderation system design.
arXiv Detail & Related papers (2024-03-09T22:50:06Z) - Toxicity Detection is NOT all you Need: Measuring the Gaps to Supporting Volunteer Content Moderators [19.401873797111662]
We conduct a model review on Hugging Face to reveal the availability of models to cover various moderation rules and guidelines.
We put state-of-the-art LLMs to the test, evaluating how well these models perform in flagging violations of platform rules from one particular forum.
Overall, we observe a non-trivial gap, as missing developed models and LLMs exhibit moderate to low performance on a significant portion of the rules.
arXiv Detail & Related papers (2023-11-14T03:18:28Z) - Why Should This Article Be Deleted? Transparent Stance Detection in
Multilingual Wikipedia Editor Discussions [47.944081120226905]
We construct a novel dataset of Wikipedia editor discussions along with their reasoning in three languages.
The dataset contains the stances of the editors (keep, delete, merge, comment), along with the stated reason, and a content moderation policy, for each edit decision.
We demonstrate that stance and corresponding reason (policy) can be predicted jointly with a high degree of accuracy, adding transparency to the decision-making process.
arXiv Detail & Related papers (2023-10-09T15:11:02Z) - Dual policy as self-model for planning [71.73710074424511]
We refer to the model used to simulate one's decisions as the agent's self-model.
Inspired by current reinforcement learning approaches and neuroscience, we explore the benefits and limitations of using a distilled policy network as the self-model.
arXiv Detail & Related papers (2023-06-07T13:58:45Z) - Explainable Abuse Detection as Intent Classification and Slot Filling [66.80201541759409]
We introduce the concept of policy-aware abuse detection, abandoning the unrealistic expectation that systems can reliably learn which phenomena constitute abuse from inspecting the data alone.
We show how architectures for intent classification and slot filling can be used for abuse detection, while providing a rationale for model decisions.
arXiv Detail & Related papers (2022-10-06T03:33:30Z) - A Trade-off-centered Framework of Content Moderation [25.068722325387515]
We find that content moderation can be characterized as a series of trade-offs around moderation actions, styles, philosophies, and values.
We argue that trade-offs should be of central importance in investigating and designing content moderation.
arXiv Detail & Related papers (2022-06-07T17:10:49Z) - Moderation Challenges in Voice-based Online Communities on Discord [24.417653462255448]
Findings suggest that the affordances of voice-based online communities change what it means to moderate content and interactions.
New ways to break rules that moderators of text-based communities find unfamiliar, such as disruptive noise and voice raiding.
New moderation strategies are limited and often based on hearsay and first impressions, resulting in problems ranging from unsuccessful moderation to false accusations.
arXiv Detail & Related papers (2021-01-13T18:43:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.