HateCheckHIn: Evaluating Hindi Hate Speech Detection Models
- URL: http://arxiv.org/abs/2205.00328v1
- Date: Sat, 30 Apr 2022 19:09:09 GMT
- Title: HateCheckHIn: Evaluating Hindi Hate Speech Detection Models
- Authors: Mithun Das and Punyajoy Saha and Binny Mathew and Animesh Mukherjee
- Abstract summary: multilingual hate is a major emerging challenge for automated detection.
We introduce a set of functionalities for the purpose of evaluation.
Considering Hindi as a base language, we craft test cases for each functionality.
- Score: 6.52974752091861
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to the sheer volume of online hate, the AI and NLP communities have
started building models to detect such hateful content. Recently, multilingual
hate is a major emerging challenge for automated detection where code-mixing or
more than one language have been used for conversation in social media.
Typically, hate speech detection models are evaluated by measuring their
performance on the held-out test data using metrics such as accuracy and
F1-score. While these metrics are useful, it becomes difficult to identify
using them where the model is failing, and how to resolve it. To enable more
targeted diagnostic insights of such multilingual hate speech models, we
introduce a set of functionalities for the purpose of evaluation. We have been
inspired to design this kind of functionalities based on real-world
conversation on social media. Considering Hindi as a base language, we craft
test cases for each functionality. We name our evaluation dataset HateCheckHIn.
To illustrate the utility of these functionalities , we test state-of-the-art
transformer based m-BERT model and the Perspective API.
Related papers
- GPT-HateCheck: Can LLMs Write Better Functional Tests for Hate Speech Detection? [50.53312866647302]
HateCheck is a suite for testing fine-grained model functionalities on synthesized data.
We propose GPT-HateCheck, a framework to generate more diverse and realistic functional tests from scratch.
Crowd-sourced annotation demonstrates that the generated test cases are of high quality.
arXiv Detail & Related papers (2024-02-23T10:02:01Z) - Evaluating ChatGPT's Performance for Multilingual and Emoji-based Hate
Speech Detection [4.809236881780707]
Large language models like ChatGPT have recently shown a great promise in performing several tasks, including hate speech detection.
This study aims to evaluate the strengths and weaknesses of the ChatGPT model in detecting hate speech at a granular level across 11 languages.
arXiv Detail & Related papers (2023-05-22T17:36:58Z) - Multilingual HateCheck: Functional Tests for Multilingual Hate Speech
Detection Models [14.128029444990895]
We introduce HateCheck (MHC), a suite of functional tests for multilingual hate speech detection models.
MHC covers 34 functionalities across ten languages, which is more languages than any other hate speech dataset.
We train and test a high-performing multilingual hate speech detection model, and reveal critical model weaknesses for monolingual and cross-lingual applications.
arXiv Detail & Related papers (2022-06-20T17:54:39Z) - A New Generation of Perspective API: Efficient Multilingual
Character-level Transformers [66.9176610388952]
We present the fundamentals behind the next version of the Perspective API from Google Jigsaw.
At the heart of the approach is a single multilingual token-free Charformer model.
We demonstrate that by forgoing static vocabularies, we gain flexibility across a variety of settings.
arXiv Detail & Related papers (2022-02-22T20:55:31Z) - Deep Learning for Hate Speech Detection: A Comparative Study [54.42226495344908]
We present here a large-scale empirical comparison of deep and shallow hate-speech detection methods.
Our goal is to illuminate progress in the area, and identify strengths and weaknesses in the current state-of-the-art.
In doing so we aim to provide guidance as to the use of hate-speech detection in practice, quantify the state-of-the-art, and identify future research directions.
arXiv Detail & Related papers (2022-02-19T03:48:20Z) - Addressing the Challenges of Cross-Lingual Hate Speech Detection [115.1352779982269]
In this paper we focus on cross-lingual transfer learning to support hate speech detection in low-resource languages.
We leverage cross-lingual word embeddings to train our neural network systems on the source language and apply it to the target language.
We investigate the issue of label imbalance of hate speech datasets, since the high ratio of non-hate examples compared to hate examples often leads to low model performance.
arXiv Detail & Related papers (2022-01-15T20:48:14Z) - Towards Language Modelling in the Speech Domain Using Sub-word
Linguistic Units [56.52704348773307]
We propose a novel LSTM-based generative speech LM based on linguistic units including syllables and phonemes.
With a limited dataset, orders of magnitude smaller than that required by contemporary generative models, our model closely approximates babbling speech.
We show the effect of training with auxiliary text LMs, multitask learning objectives, and auxiliary articulatory features.
arXiv Detail & Related papers (2021-10-31T22:48:30Z) - Cross-lingual hate speech detection based on multilingual
domain-specific word embeddings [4.769747792846004]
We propose to address the problem of multilingual hate speech detection from the perspective of transfer learning.
Our goal is to determine if knowledge from one particular language can be used to classify other language.
We show that the use of our simple yet specific multilingual hate representations improves classification results.
arXiv Detail & Related papers (2021-04-30T02:24:50Z) - HateCheck: Functional Tests for Hate Speech Detection Models [3.4938484663205776]
We introduce HateCheck, a first suite of functional tests for hate speech detection models.
We specify 29 model functionalities, the selection of which we motivate by reviewing previous research.
We test near-state-of-the-art transformer detection models as well as a popular commercial model, revealing critical model weaknesses.
arXiv Detail & Related papers (2020-12-31T13:44:56Z) - Comparison of Interactive Knowledge Base Spelling Correction Models for
Low-Resource Languages [81.90356787324481]
Spelling normalization for low resource languages is a challenging task because the patterns are hard to predict.
This work shows a comparison of a neural model and character language models with varying amounts on target language data.
Our usage scenario is interactive correction with nearly zero amounts of training examples, improving models as more data is collected.
arXiv Detail & Related papers (2020-10-20T17:31:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.