K-HATERS: A Hate Speech Detection Corpus in Korean with Target-Specific
Ratings
- URL: http://arxiv.org/abs/2310.15439v1
- Date: Tue, 24 Oct 2023 01:20:05 GMT
- Title: K-HATERS: A Hate Speech Detection Corpus in Korean with Target-Specific
Ratings
- Authors: Chaewon Park, Soohwan Kim, Kyubyong Park, Kunwoo Park
- Abstract summary: K-HATERS is a new corpus for hate speech detection in Korean, comprising approximately 192K news comments with target-specific offensiveness ratings.
This study contributes to the NLP research on hate speech detection and resource construction.
- Score: 6.902524826065157
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Numerous datasets have been proposed to combat the spread of online hate.
Despite these efforts, a majority of these resources are English-centric,
primarily focusing on overt forms of hate. This research gap calls for
developing high-quality corpora in diverse languages that also encapsulate more
subtle hate expressions. This study introduces K-HATERS, a new corpus for hate
speech detection in Korean, comprising approximately 192K news comments with
target-specific offensiveness ratings. This resource is the largest offensive
language corpus in Korean and is the first to offer target-specific ratings on
a three-point Likert scale, enabling the detection of hate expressions in
Korean across varying degrees of offensiveness. We conduct experiments showing
the effectiveness of the proposed corpus, including a comparison with existing
datasets. Additionally, to address potential noise and bias in human
annotations, we explore a novel idea of adopting the Cognitive Reflection Test,
which is widely used in social science for assessing an individual's cognitive
ability, as a proxy of labeling quality. Findings indicate that annotations
from individuals with the lowest test scores tend to yield detection models
that make biased predictions toward specific target groups and are less
accurate. This study contributes to the NLP research on hate speech detection
and resource construction. The code and dataset can be accessed at
https://github.com/ssu-humane/K-HATERS.
Related papers
- HCDIR: End-to-end Hate Context Detection, and Intensity Reduction model
for online comments [2.162419921663162]
We propose a novel end-to-end model, HCDIR, for Hate Context Detection, and Hate Intensity Reduction in social media posts.
We fine-tuned several pre-trained language models to detect hateful comments to ascertain the best-performing hateful comments detection model.
arXiv Detail & Related papers (2023-12-20T17:05:46Z) - SOUL: Towards Sentiment and Opinion Understanding of Language [96.74878032417054]
We propose a new task called Sentiment and Opinion Understanding of Language (SOUL)
SOUL aims to evaluate sentiment understanding through two subtasks: Review (RC) and Justification Generation (JG)
arXiv Detail & Related papers (2023-10-27T06:48:48Z) - Quantifying the Dialect Gap and its Correlates Across Languages [69.18461982439031]
This work will lay the foundation for furthering the field of dialectal NLP by laying out evident disparities and identifying possible pathways for addressing them through mindful data collection.
arXiv Detail & Related papers (2023-10-23T17:42:01Z) - K-MHaS: A Multi-label Hate Speech Detection Dataset in Korean Online
News Comment [3.428320237347854]
We introduce K-MHaS, a new multi-label dataset for hate speech detection that effectively handles Korean language patterns.
The dataset consists of 109k utterances from news comments and provides multi-label classification from 1 to 4 labels.
KR-BERT with sub-character tokenizer outperforms, recognising decomposed characters in each hate speech class.
arXiv Detail & Related papers (2022-08-23T02:10:53Z) - KOLD: Korean Offensive Language Dataset [11.699797031874233]
We present a Korean offensive language dataset (KOLD), 40k comments labeled with offensiveness, target, and targeted group information.
We show that title information serves as context and is helpful to discern the target of hatred, especially when they are omitted in the comment.
arXiv Detail & Related papers (2022-05-23T13:58:45Z) - APEACH: Attacking Pejorative Expressions with Analysis on
Crowd-Generated Hate Speech Evaluation Datasets [4.034948808542701]
APEACH is a method that allows the collection of hate speech generated by unspecified users.
By controlling the crowd-generation of hate speech and adding only a minimum post-labeling, we create a corpus that enables the generalizable and fair evaluation of hate speech detection.
arXiv Detail & Related papers (2022-02-25T02:04:38Z) - Deep Learning for Hate Speech Detection: A Comparative Study [54.42226495344908]
We present here a large-scale empirical comparison of deep and shallow hate-speech detection methods.
Our goal is to illuminate progress in the area, and identify strengths and weaknesses in the current state-of-the-art.
In doing so we aim to provide guidance as to the use of hate-speech detection in practice, quantify the state-of-the-art, and identify future research directions.
arXiv Detail & Related papers (2022-02-19T03:48:20Z) - COLD: A Benchmark for Chinese Offensive Language Detection [54.60909500459201]
We use COLDataset, a Chinese offensive language dataset with 37k annotated sentences.
We also propose textscCOLDetector to study output offensiveness of popular Chinese language models.
Our resources and analyses are intended to help detoxify the Chinese online communities and evaluate the safety performance of generative language models.
arXiv Detail & Related papers (2022-01-16T11:47:23Z) - Addressing the Challenges of Cross-Lingual Hate Speech Detection [115.1352779982269]
In this paper we focus on cross-lingual transfer learning to support hate speech detection in low-resource languages.
We leverage cross-lingual word embeddings to train our neural network systems on the source language and apply it to the target language.
We investigate the issue of label imbalance of hate speech datasets, since the high ratio of non-hate examples compared to hate examples often leads to low model performance.
arXiv Detail & Related papers (2022-01-15T20:48:14Z) - AM2iCo: Evaluating Word Meaning in Context across Low-ResourceLanguages
with Adversarial Examples [51.048234591165155]
We present AM2iCo, Adversarial and Multilingual Meaning in Context.
It aims to faithfully assess the ability of state-of-the-art (SotA) representation models to understand the identity of word meaning in cross-lingual contexts.
Results reveal that current SotA pretrained encoders substantially lag behind human performance.
arXiv Detail & Related papers (2021-04-17T20:23:45Z) - BEEP! Korean Corpus of Online News Comments for Toxic Speech Detection [3.90603670603335]
We first present 9.4K manually labeled entertainment news comments for identifying Korean toxic speech.
The comments are annotated regarding social bias and hate speech since both aspects are correlated.
We provide benchmarks using CharCNN, BiLSTM, and BERT, where BERT achieves the highest score on all tasks.
arXiv Detail & Related papers (2020-05-26T03:34:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.