Explain Thyself Bully: Sentiment Aided Cyberbullying Detection with
Explanation
- URL: http://arxiv.org/abs/2401.09023v1
- Date: Wed, 17 Jan 2024 07:36:22 GMT
- Title: Explain Thyself Bully: Sentiment Aided Cyberbullying Detection with
Explanation
- Authors: Krishanu Maity, Prince Jha, Raghav Jain, Sriparna Saha, Pushpak
Bhattacharyya
- Abstract summary: Cyberbullying has become a big issue with the popularity of different social media networks and online communication apps.
Recent laws like "right to explanations" of General Data Protection Regulation have spurred research in developing interpretable models.
We develop first interpretable multi-task model called em mExCB for automatic cyberbullying detection from code-mixed languages.
- Score: 52.3781496277104
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Cyberbullying has become a big issue with the popularity of different social
media networks and online communication apps. While plenty of research is going
on to develop better models for cyberbullying detection in monolingual
language, there is very little research on the code-mixed languages and
explainability aspect of cyberbullying. Recent laws like "right to
explanations" of General Data Protection Regulation, have spurred research in
developing interpretable models rather than focusing on performance. Motivated
by this we develop the first interpretable multi-task model called {\em mExCB}
for automatic cyberbullying detection from code-mixed languages which can
simultaneously solve several tasks, cyberbullying detection,
explanation/rationale identification, target group detection and sentiment
analysis. We have introduced {\em BullyExplain}, the first benchmark dataset
for explainable cyberbullying detection in code-mixed language. Each post in
{\em BullyExplain} dataset is annotated with four labels, i.e., {\em bully
label, sentiment label, target and rationales (explainability)}, i.e., which
phrases are being responsible for annotating the post as a bully. The proposed
multitask framework (mExCB) based on CNN and GRU with word and sub-sentence
(SS) level attention is able to outperform several baselines and state of the
art models when applied on {\em BullyExplain} dataset.
Related papers
- The Use of a Large Language Model for Cyberbullying Detection [0.0]
cyberbullying (CB) is the most prevalent phenomenon in todays cyber world.
It is a severe threat to the mental and physical health of citizens.
This opens the need to develop a robust system to prevent bullying content from online forums, blogs, and social media platforms.
arXiv Detail & Related papers (2024-02-06T15:46:31Z) - Meme-ingful Analysis: Enhanced Understanding of Cyberbullying in Memes
Through Multimodal Explanations [48.82168723932981]
We introduce em MultiBully-Ex, the first benchmark dataset for multimodal explanation from code-mixed cyberbullying memes.
A Contrastive Language-Image Pretraining (CLIP) approach has been proposed for visual and textual explanation of a meme.
arXiv Detail & Related papers (2024-01-18T11:24:30Z) - Cyberbullying Detection for Low-resource Languages and Dialects: Review
of the State of the Art [0.9831489366502298]
There are 23 low-resource languages and dialects covered by this paper, including Bangla, Hindi, Dravidian languages and others.
In the survey, we identify some of the research gaps of previous studies, which include the lack of reliable definitions of cyberbullying.
Based on those proposed suggestions, we collect and release a cyberbullying dataset in the Chittagonian dialect of Bangla.
arXiv Detail & Related papers (2023-08-30T03:52:28Z) - How to Solve Few-Shot Abusive Content Detection Using the Data We Actually Have [58.23138483086277]
In this work we leverage datasets we already have, covering a wide range of tasks related to abusive language detection.
Our goal is to build models cheaply for a new target label set and/or language, using only a few training examples of the target domain.
Our experiments show that using already existing datasets and only a few-shots of the target task the performance of models improve both monolingually and across languages.
arXiv Detail & Related papers (2023-05-23T14:04:12Z) - Verifying the Robustness of Automatic Credibility Assessment [79.08422736721764]
Text classification methods have been widely investigated as a way to detect content of low credibility.
In some cases insignificant changes in input text can mislead the models.
We introduce BODEGA: a benchmark for testing both victim models and attack methods on misinformation detection tasks.
arXiv Detail & Related papers (2023-03-14T16:11:47Z) - Addressing the Challenges of Cross-Lingual Hate Speech Detection [115.1352779982269]
In this paper we focus on cross-lingual transfer learning to support hate speech detection in low-resource languages.
We leverage cross-lingual word embeddings to train our neural network systems on the source language and apply it to the target language.
We investigate the issue of label imbalance of hate speech datasets, since the high ratio of non-hate examples compared to hate examples often leads to low model performance.
arXiv Detail & Related papers (2022-01-15T20:48:14Z) - Pseudo Siamese Network for Few-shot Intent Generation [54.10596778418007]
We propose a Pseudo Siamese Network (PSN) to generate labeled data for few-shot intents.
PSN consists of two identicalworks with the same structure but different weights: an action network and an object network.
Experiments on two real-world datasets show that PSN achieves state-of-the-art performance for the generalized few shot intent detection task.
arXiv Detail & Related papers (2021-05-03T14:30:47Z) - Analysing Cyberbullying using Natural Language Processing by
Understanding Jargon in Social Media [4.932130498861987]
In our work, we explore binary classification by using a combination of datasets from various social media platforms.
We experiment through multiple models such as Bi-LSTM, GloVe, state-of-the-art models like BERT, and apply a unique preprocessing technique by introducing a slang-abusive corpus.
arXiv Detail & Related papers (2021-04-23T04:20:19Z) - HENIN: Learning Heterogeneous Neural Interaction Networks for
Explainable Cyberbullying Detection on Social Media [11.443698975923176]
We propose a novel deep model, HEterogeneous Neural Interaction Networks (HENIN), for explainable cyberbullying detection.
HENIN contains the following components: a comment encoder, a post-comment co-attention sub-network, and session-session and post-post interaction extractors.
arXiv Detail & Related papers (2020-10-09T13:44:34Z) - Aggressive, Repetitive, Intentional, Visible, and Imbalanced: Refining
Representations for Cyberbullying Classification [4.945634077636197]
We study the nuanced problem of cyberbullying using five explicit factors to represent its social and linguistic aspects.
These results demonstrate the importance of representing and modeling cyberbullying as a social phenomenon.
arXiv Detail & Related papers (2020-04-04T00:35:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.