Echoes of Discord: Forecasting Hater Reactions to Counterspeech
- URL: http://arxiv.org/abs/2501.16235v2
- Date: Thu, 13 Feb 2025 20:14:27 GMT
- Title: Echoes of Discord: Forecasting Hater Reactions to Counterspeech
- Authors: Xiaoying Song, Sharon Lisseth Perez, Xinchen Yu, Eduardo Blanco, Lingzi Hong,
- Abstract summary: This study analyzes the impact of counterspeech from the hater's perspective.
We employ two strategies: a two-stage reaction predictor and a three-way classification model.
Experimental results demonstrate that the 3-way classification model outperforms the two-stage reaction predictor.
- Score: 10.658005418397748
- License:
- Abstract: Hate speech (HS) erodes the inclusiveness of online users and propagates negativity and division. Counterspeech has been recognized as a way to mitigate the harmful consequences. While some research has investigated the impact of user-generated counterspeech on social media platforms, few have examined and modeled haters' reactions toward counterspeech, despite the immediate alteration of haters' attitudes being an important aspect of counterspeech. This study fills the gap by analyzing the impact of counterspeech from the hater's perspective, focusing on whether the counterspeech leads the hater to reenter the conversation and if the reentry is hateful. We compile the Reddit Echoes of Hate dataset (ReEco), which consists of triple-turn conversations featuring haters' reactions, to assess the impact of counterspeech. To predict haters' behaviors, we employ two strategies: a two-stage reaction predictor and a three-way classifier. The linguistic analysis sheds insights on the language of counterspeech to hate eliciting different haters' reactions. Experimental results demonstrate that the 3-way classification model outperforms the two-stage reaction predictor, which first predicts reentry and then determines the reentry type. We conclude the study with an assessment showing the most common errors identified by the best-performing model.
Related papers
- CounterQuill: Investigating the Potential of Human-AI Collaboration in Online Counterspeech Writing [6.929003593008481]
We introduce CounterQuill, an AI-mediated system that assists users in composing effective and empathetic counterspeech.
CounterQuill provides a three-step process: (1) a learning session to help users understand hate speech and counterspeech; (2) a brainstorming session that guides users in identifying key elements of hate speech and exploring counterspeech strategies; and (3) a co-writing session that enables users to draft and refine their counterspeech with CounterQuill.
arXiv Detail & Related papers (2024-10-03T22:29:20Z) - Hostile Counterspeech Drives Users From Hate Subreddits [1.5035331281822]
We analyze the effect of counterspeech on newcomers within hate subreddits on Reddit.
Non-hostile counterspeech is ineffective at keeping users from fully disengaging from these hate subreddits.
A single hostile counterspeech comment substantially reduces both future likelihood of engagement.
arXiv Detail & Related papers (2024-05-28T17:12:41Z) - Behind the Counter: Exploring the Motivations and Barriers of Online Counterspeech Writing [6.790819952175892]
Having been a target of online hate is a key driver of frequent online counterspeech engagement.
People differ in their motivations and barriers towards engaging in online counterspeech across different demographic groups.
arXiv Detail & Related papers (2024-03-25T18:56:35Z) - Beyond Denouncing Hate: Strategies for Countering Implied Biases and
Stereotypes in Language [18.560379338032558]
We draw from psychology and philosophy literature to craft six psychologically inspired strategies to challenge the underlying stereotypical implications of hateful language.
We show that human-written counterspeech uses strategies that are more specific to the implied stereotype, whereas machine-generated counterspeech uses less specific strategies.
Our findings point to the importance of accounting for the underlying stereotypical implications of speech when generating counterspeech.
arXiv Detail & Related papers (2023-10-31T21:33:46Z) - Measuring the Effect of Influential Messages on Varying Personas [67.1149173905004]
We present a new task, Response Forecasting on Personas for News Media, to estimate the response a persona might have upon seeing a news message.
The proposed task not only introduces personalization in the modeling but also predicts the sentiment polarity and intensity of each response.
This enables more accurate and comprehensive inference on the mental state of the persona.
arXiv Detail & Related papers (2023-05-25T21:01:00Z) - ReactFace: Online Multiple Appropriate Facial Reaction Generation in Dyadic Interactions [46.66378299720377]
In dyadic interaction, predicting the listener's facial reactions is challenging as different reactions could be appropriate in response to the same speaker's behaviour.
This paper reformulates the task as an extrapolation or prediction problem, and proposes a novel framework (called ReactFace) to generate multiple different but appropriate facial reactions.
arXiv Detail & Related papers (2023-05-25T05:55:53Z) - CoSyn: Detecting Implicit Hate Speech in Online Conversations Using a
Context Synergized Hyperbolic Network [52.85130555886915]
CoSyn is a context-synergized neural network that explicitly incorporates user- and conversational context for detecting implicit hate speech in online conversations.
We show that CoSyn outperforms all our baselines in detecting implicit hate speech with absolute improvements in the range of 1.24% - 57.8%.
arXiv Detail & Related papers (2023-03-02T17:30:43Z) - Addressing the Challenges of Cross-Lingual Hate Speech Detection [115.1352779982269]
In this paper we focus on cross-lingual transfer learning to support hate speech detection in low-resource languages.
We leverage cross-lingual word embeddings to train our neural network systems on the source language and apply it to the target language.
We investigate the issue of label imbalance of hate speech datasets, since the high ratio of non-hate examples compared to hate examples often leads to low model performance.
arXiv Detail & Related papers (2022-01-15T20:48:14Z) - Characterizing the adversarial vulnerability of speech self-supervised
learning [95.03389072594243]
We make the first attempt to investigate the adversarial vulnerability of such paradigm under the attacks from both zero-knowledge adversaries and limited-knowledge adversaries.
The experimental results illustrate that the paradigm proposed by SUPERB is seriously vulnerable to limited-knowledge adversaries.
arXiv Detail & Related papers (2021-11-08T08:44:04Z) - Hate Speech Classifiers Learn Human-Like Social Stereotypes [4.132204773132937]
Social stereotypes negatively impact individuals' judgements about different groups.
Social stereotypes may have a critical role in how people understand language directed toward minority social groups.
arXiv Detail & Related papers (2021-10-28T01:35:41Z) - Constructing interval variables via faceted Rasch measurement and
multitask deep learning: a hate speech application [63.10266319378212]
We propose a method for measuring complex variables on a continuous, interval spectrum by combining supervised deep learning with the Constructing Measures approach to faceted Rasch item response theory (IRT)
We demonstrate this new method on a dataset of 50,000 social media comments sourced from YouTube, Twitter, and Reddit and labeled by 11,000 U.S.-based Amazon Mechanical Turk workers.
arXiv Detail & Related papers (2020-09-22T02:15:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.