Causal Understanding of Why Users Share Hate Speech on Social Media
- URL: http://arxiv.org/abs/2310.15772v1
- Date: Tue, 24 Oct 2023 12:17:48 GMT
- Title: Causal Understanding of Why Users Share Hate Speech on Social Media
- Authors: Dominique Geissler and Abdurahman Maarouf and Stefan Feuerriegel
- Abstract summary: We present a comprehensive, causal analysis of the user attributes that make users reshare hate speech.
We develop a novel, three-step causal framework.
We find that users with fewer followers, fewer friends, and fewer posts share more hate speech.
- Score: 23.128421664169654
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Hate speech on social media threatens the mental and physical well-being of
individuals and is further responsible for real-world violence. An important
driver behind the spread of hate speech and thus why hateful posts can go viral
are reshares, yet little is known about why users reshare hate speech. In this
paper, we present a comprehensive, causal analysis of the user attributes that
make users reshare hate speech. However, causal inference from observational
social media data is challenging, because such data likely suffer from
selection bias, and there is further confounding due to differences in the
vulnerability of users to hate speech. We develop a novel, three-step causal
framework: (1) We debias the observational social media data by applying
inverse propensity scoring. (2) We use the debiased propensity scores to model
the latent vulnerability of users to hate speech as a latent embedding. (3) We
model the causal effects of user attributes on users' probability of sharing
hate speech, while controlling for the latent vulnerability of users to hate
speech. Compared to existing baselines, a particular strength of our framework
is that it models causal effects that are non-linear, yet still explainable. We
find that users with fewer followers, fewer friends, and fewer posts share more
hate speech. Younger accounts, in return, share less hate speech. Overall,
understanding the factors that drive users to share hate speech is crucial for
detecting individuals at risk of engaging in harmful behavior and for designing
effective mitigation strategies.
Related papers
- All You Need is "Leet": Evading Hate-speech Detection AI [0.6906005491572401]
In this paper, we design black-box techniques to protect users from hate-speech on online platforms.<n>Our best perturbation attack is successfully able to evade hate-speech detection for 86.8 % of hateful text.
arXiv Detail & Related papers (2025-05-22T05:55:26Z) - A Federated Approach to Few-Shot Hate Speech Detection for Marginalized Communities [43.37824420609252]
Hate speech online remains an understudied issue for marginalized communities.<n>In this paper, we aim to provide marginalized communities with a privacy-preserving tool to protect themselves from online hate speech.
arXiv Detail & Related papers (2024-12-06T11:00:05Z) - ProvocationProbe: Instigating Hate Speech Dataset from Twitter [0.39052860539161904]
textitProvocationProbe is a dataset designed to explore what distinguishes instigating hate speech from general hate speech.
For this study, we collected around twenty thousand tweets from Twitter, encompassing a total of nine global controversies.
arXiv Detail & Related papers (2024-10-25T16:57:59Z) - Hostile Counterspeech Drives Users From Hate Subreddits [1.5035331281822]
We analyze the effect of counterspeech on newcomers within hate subreddits on Reddit.
Non-hostile counterspeech is ineffective at keeping users from fully disengaging from these hate subreddits.
A single hostile counterspeech comment substantially reduces both future likelihood of engagement.
arXiv Detail & Related papers (2024-05-28T17:12:41Z) - An Investigation of Large Language Models for Real-World Hate Speech
Detection [46.15140831710683]
A major limitation of existing methods is that hate speech detection is a highly contextual problem.
Recently, large language models (LLMs) have demonstrated state-of-the-art performance in several natural language tasks.
Our study reveals that a meticulously crafted reasoning prompt can effectively capture the context of hate speech.
arXiv Detail & Related papers (2024-01-07T00:39:33Z) - On the rise of fear speech in online social media [7.090807766284268]
Fear speech, as the name suggests, attempts to incite fear about a target community.
This article presents a large-scale study to understand the prevalence of 400K fear speech and over 700K hate speech posts collected from Gab.com.
arXiv Detail & Related papers (2023-03-18T02:46:49Z) - CoSyn: Detecting Implicit Hate Speech in Online Conversations Using a
Context Synergized Hyperbolic Network [52.85130555886915]
CoSyn is a context-synergized neural network that explicitly incorporates user- and conversational context for detecting implicit hate speech in online conversations.
We show that CoSyn outperforms all our baselines in detecting implicit hate speech with absolute improvements in the range of 1.24% - 57.8%.
arXiv Detail & Related papers (2023-03-02T17:30:43Z) - Quantifying How Hateful Communities Radicalize Online Users [2.378428291297535]
We measure the impact of joining fringe hateful communities in terms of hate speech propagated to the rest of the social network.
We use data from Reddit to assess the effect of joining one type of echo chamber: a digital community of like-minded users exhibiting hateful behavior.
We show that the harmful speech does not remain contained within the community.
arXiv Detail & Related papers (2022-09-19T01:13:29Z) - Addressing the Challenges of Cross-Lingual Hate Speech Detection [115.1352779982269]
In this paper we focus on cross-lingual transfer learning to support hate speech detection in low-resource languages.
We leverage cross-lingual word embeddings to train our neural network systems on the source language and apply it to the target language.
We investigate the issue of label imbalance of hate speech datasets, since the high ratio of non-hate examples compared to hate examples often leads to low model performance.
arXiv Detail & Related papers (2022-01-15T20:48:14Z) - Nipping in the Bud: Detection, Diffusion and Mitigation of Hate Speech
on Social Media [21.47216483704825]
This article presents methodological challenges that hinder building automated hate mitigation systems.
We discuss a series of our proposed solutions to limit the spread of hate speech on social media.
arXiv Detail & Related papers (2022-01-04T03:44:46Z) - Annotators with Attitudes: How Annotator Beliefs And Identities Bias
Toxic Language Detection [75.54119209776894]
We investigate the effect of annotator identities (who) and beliefs (why) on toxic language annotations.
We consider posts with three characteristics: anti-Black language, African American English dialect, and vulgarity.
Our results show strong associations between annotator identity and beliefs and their ratings of toxicity.
arXiv Detail & Related papers (2021-11-15T18:58:20Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - Racism is a Virus: Anti-Asian Hate and Counterspeech in Social Media
during the COVID-19 Crisis [51.39895377836919]
COVID-19 has sparked racism and hate on social media targeted towards Asian communities.
We study the evolution and spread of anti-Asian hate speech through the lens of Twitter.
We create COVID-HATE, the largest dataset of anti-Asian hate and counterspeech spanning 14 months.
arXiv Detail & Related papers (2020-05-25T21:58:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.