Automatic Expansion and Retargeting of Arabic Offensive Language
Training
- URL: http://arxiv.org/abs/2111.09574v1
- Date: Thu, 18 Nov 2021 08:25:09 GMT
- Title: Automatic Expansion and Retargeting of Arabic Offensive Language
Training
- Authors: Hamdy Mubarak, Ahmed Abdelali, Kareem Darwish and Younes Samih
- Abstract summary: We employ two key insights, namely that replies on Twitter often imply opposition and some accounts are persistent in their offensiveness towards specific targets.
We show the efficacy of the approach on Arabic tweets with 13% and 79% relative F1-measure improvement in entity specific offensive language detection.
- Score: 12.111859709582617
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Rampant use of offensive language on social media led to recent efforts on
automatic identification of such language. Though offensive language has
general characteristics, attacks on specific entities may exhibit distinct
phenomena such as malicious alterations in the spelling of names. In this
paper, we present a method for identifying entity specific offensive language.
We employ two key insights, namely that replies on Twitter often imply
opposition and some accounts are persistent in their offensiveness towards
specific targets. Using our methodology, we are able to collect thousands of
targeted offensive tweets. We show the efficacy of the approach on Arabic
tweets with 13% and 79% relative F1-measure improvement in entity specific
offensive language detection when using deep-learning based and support vector
machine based classifiers respectively. Further, expanding the training set
with automatically identified offensive tweets directed at multiple entities
can improve F1-measure by 48%.
Related papers
- Muted: Multilingual Targeted Offensive Speech Identification and
Visualization [15.656203119337436]
Muted is a system to identify multilingual HAP content by displaying offensive arguments and their targets using heat maps to indicate their intensity.
We present the model's performance on identifying offensive spans and their targets in existing datasets and present new annotations on German text.
arXiv Detail & Related papers (2023-12-18T16:50:27Z) - Locally Differentially Private Document Generation Using Zero Shot
Prompting [61.20953109732442]
We propose a locally differentially private mechanism called DP-Prompt to counter author de-anonymization attacks.
When DP-Prompt is used with a powerful language model like ChatGPT (gpt-3.5), we observe a notable reduction in the success rate of de-anonymization attacks.
arXiv Detail & Related papers (2023-10-24T18:25:13Z) - Unlikelihood Tuning on Negative Samples Amazingly Improves Zero-Shot
Translation [79.96416609433724]
Zero-shot translation (ZST) aims to translate between unseen language pairs in training data.
The common practice to guide the zero-shot language mapping during inference is to deliberately insert the source and target language IDs.
Recent studies have shown that language IDs sometimes fail to navigate the ZST task, making them suffer from the off-target problem.
arXiv Detail & Related papers (2023-09-28T17:02:36Z) - How to Solve Few-Shot Abusive Content Detection Using the Data We Actually Have [58.23138483086277]
In this work we leverage datasets we already have, covering a wide range of tasks related to abusive language detection.
Our goal is to build models cheaply for a new target label set and/or language, using only a few training examples of the target domain.
Our experiments show that using already existing datasets and only a few-shots of the target task the performance of models improve both monolingually and across languages.
arXiv Detail & Related papers (2023-05-23T14:04:12Z) - On The Robustness of Offensive Language Classifiers [10.742675209112623]
Social media platforms are deploying machine learning based offensive language classification systems to combat hateful, racist, and other forms of offensive speech at scale.
We study the robustness of state-of-the-art offensive language classifiers against more crafty adversarial attacks.
Our results show that these crafty adversarial attacks can degrade the accuracy of offensive language classifiers by more than 50% while also being able to preserve the readability and meaning of the modified text.
arXiv Detail & Related papers (2022-03-21T20:44:30Z) - Emojis as Anchors to Detect Arabic Offensive Language and Hate Speech [6.1875341699258595]
We introduce a generic, language-independent method to collect a large percentage of offensive and hate tweets.
We harness the extralinguistic information embedded in the emojis to collect a large number of offensive tweets.
arXiv Detail & Related papers (2022-01-18T03:56:57Z) - COLD: A Benchmark for Chinese Offensive Language Detection [54.60909500459201]
We use COLDataset, a Chinese offensive language dataset with 37k annotated sentences.
We also propose textscCOLDetector to study output offensiveness of popular Chinese language models.
Our resources and analyses are intended to help detoxify the Chinese online communities and evaluate the safety performance of generative language models.
arXiv Detail & Related papers (2022-01-16T11:47:23Z) - Leveraging Adversarial Training in Self-Learning for Cross-Lingual Text
Classification [52.69730591919885]
We present a semi-supervised adversarial training process that minimizes the maximal loss for label-preserving input perturbations.
We observe significant gains in effectiveness on document and intent classification for a diverse set of languages.
arXiv Detail & Related papers (2020-07-29T19:38:35Z) - Kungfupanda at SemEval-2020 Task 12: BERT-Based Multi-Task Learning for
Offensive Language Detection [55.445023584632175]
We build an offensive language detection system, which combines multi-task learning with BERT-based models.
Our model achieves 91.51% F1 score in English Sub-task A, which is comparable to the first place.
arXiv Detail & Related papers (2020-04-28T11:27:24Z) - Arabic Offensive Language on Twitter: Analysis and Experiments [9.879488163141813]
We introduce a method for building a dataset that is not biased by topic, dialect, or target.
We produce the largest Arabic dataset to date with special tags for vulgarity and hate speech.
arXiv Detail & Related papers (2020-04-05T13:05:11Z) - Offensive Language Detection: A Comparative Analysis [2.5739449801033842]
We explore the effectiveness of Google sentence encoder, Fasttext, Dynamic mode decomposition (DMD) based features and Random kitchen sink (RKS) method for offensive language detection.
From the experiments and evaluation we observed that RKS with fastetxt achieved competing results.
arXiv Detail & Related papers (2020-01-09T17:48:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.