Detecting East Asian Prejudice on Social Media
- URL: http://arxiv.org/abs/2005.03909v1
- Date: Fri, 8 May 2020 08:53:47 GMT
- Title: Detecting East Asian Prejudice on Social Media
- Authors: Bertie Vidgen, Austin Botelho, David Broniatowski, Ella Guest, Matthew
Hall, Helen Margetts, Rebekah Tromble, Zeerak Waseem, Scott Hale
- Abstract summary: We report on the creation of a classifier that detects and categorizes social media posts from Twitter into four classes: Hostility against East Asia, Criticism of East Asia, Meta-discussions of East Asian prejudice and a neutral class.
- Score: 10.647940201343575
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The outbreak of COVID-19 has transformed societies across the world as
governments tackle the health, economic and social costs of the pandemic. It
has also raised concerns about the spread of hateful language and prejudice
online, especially hostility directed against East Asia. In this paper we
report on the creation of a classifier that detects and categorizes social
media posts from Twitter into four classes: Hostility against East Asia,
Criticism of East Asia, Meta-discussions of East Asian prejudice and a neutral
class. The classifier achieves an F1 score of 0.83 across all four classes. We
provide our final model (coded in Python), as well as a new 20,000 tweet
training dataset used to make the classifier, two analyses of hashtags
associated with East Asian prejudice and the annotation codebook. The
classifier can be implemented by other researchers, assisting with both online
content moderation processes and further research into the dynamics, prevalence
and impact of East Asian prejudice online during this global pandemic.
Related papers
- SPEED++: A Multilingual Event Extraction Framework for Epidemic Prediction and Preparedness [73.73883111570458]
We introduce the first multilingual Event Extraction framework for extracting epidemic event information for a wide range of diseases and languages.
Annotating data in every language is infeasible; thus we develop zero-shot cross-lingual cross-disease models.
Our framework can provide epidemic warnings for COVID-19 in its earliest stages in Dec 2019 from Chinese Weibo posts without any training in Chinese.
arXiv Detail & Related papers (2024-10-24T03:03:54Z) - A Community-Centric Perspective for Characterizing and Detecting Anti-Asian Violence-Provoking Speech [31.98433210638392]
Violence-provoking speech contributed to a massive surge in anti-Asian crimes during the pandemic.
We develop a codebook to characterize anti-Asian violence-provoking speech and collect a community-sourced dataset.
We show that accurate and reliable detection of violence-provoking speech is a challenging task.
arXiv Detail & Related papers (2024-07-21T17:27:17Z) - Are Personalized Stochastic Parrots More Dangerous? Evaluating Persona
Biases in Dialogue Systems [103.416202777731]
We study "persona biases", which we define to be the sensitivity of dialogue models' harmful behaviors contingent upon the personas they adopt.
We categorize persona biases into biases in harmful expression and harmful agreement, and establish a comprehensive evaluation framework to measure persona biases in five aspects: Offensiveness, Toxic Continuation, Regard, Stereotype Agreement, and Toxic Agreement.
arXiv Detail & Related papers (2023-10-08T21:03:18Z) - KoSBi: A Dataset for Mitigating Social Bias Risks Towards Safer Large
Language Model Application [45.3863281375947]
Large language models (LLMs) learn not only natural text generation abilities but also social biases against different demographic groups from real-world data.
Existing research and resources are not readily applicable in South Korea due to the differences in language and culture.
We present KO SB I, a new social bias dataset of 34k pairs of contexts and sentences in Korean covering 72 demographic groups in 15 categories.
arXiv Detail & Related papers (2023-05-28T12:07:16Z) - Multi-dimensional Racism Classification during COVID-19: Stigmatization,
Offensiveness, Blame, and Exclusion [1.0878040851638]
We develop a multi-dimensional model for racism detection, namely stigmatization, offensiveness, blame, and exclusion.
This categorical detection enables insights into the underlying subtlety of racist discussion on digital platforms during COVID-19.
arXiv Detail & Related papers (2022-08-29T00:38:56Z) - "Stop Asian Hate!" : Refining Detection of Anti-Asian Hate Speech During
the COVID-19 Pandemic [2.5227595609842206]
COVID-19 pandemic has fueled a surge in anti-Asian xenophobia and prejudice.
We create and annotate a corpus of Twitter tweets using 2 experimental approaches to explore anti-Asian abusive and hate speech.
arXiv Detail & Related papers (2021-12-04T06:55:19Z) - Predicting Anti-Asian Hateful Users on Twitter during COVID-19 [7.788173128266611]
We apply natural language processing techniques to characterize social media users who began to post anti-Asian hate messages during COVID-19.
It is possible to predict who later publicly posted anti-Asian slurs.
arXiv Detail & Related papers (2021-09-15T13:49:37Z) - Towards Understanding and Mitigating Social Biases in Language Models [107.82654101403264]
Large-scale pretrained language models (LMs) can be potentially dangerous in manifesting undesirable representational biases.
We propose steps towards mitigating social biases during text generation.
Our empirical results and human evaluation demonstrate effectiveness in mitigating bias while retaining crucial contextual information.
arXiv Detail & Related papers (2021-06-24T17:52:43Z) - Revealing Persona Biases in Dialogue Systems [64.96908171646808]
We present the first large-scale study on persona biases in dialogue systems.
We conduct analyses on personas of different social classes, sexual orientations, races, and genders.
In our studies of the Blender and DialoGPT dialogue systems, we show that the choice of personas can affect the degree of harms in generated responses.
arXiv Detail & Related papers (2021-04-18T05:44:41Z) - Hate Speech Detection and Racial Bias Mitigation in Social Media based
on BERT model [1.9336815376402716]
We introduce a transfer learning approach for hate speech detection based on an existing pre-trained language model called BERT.
We evaluate the proposed model on two publicly available datasets annotated for racism, sexism, hate or offensive content on Twitter.
arXiv Detail & Related papers (2020-08-14T16:47:25Z) - Racism is a Virus: Anti-Asian Hate and Counterspeech in Social Media
during the COVID-19 Crisis [51.39895377836919]
COVID-19 has sparked racism and hate on social media targeted towards Asian communities.
We study the evolution and spread of anti-Asian hate speech through the lens of Twitter.
We create COVID-HATE, the largest dataset of anti-Asian hate and counterspeech spanning 14 months.
arXiv Detail & Related papers (2020-05-25T21:58:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.