Explain Thyself Bully: Sentiment Aided Cyberbullying Detection with
Explanation
- URL: http://arxiv.org/abs/2401.09023v1
- Date: Wed, 17 Jan 2024 07:36:22 GMT
- Title: Explain Thyself Bully: Sentiment Aided Cyberbullying Detection with
Explanation
- Authors: Krishanu Maity, Prince Jha, Raghav Jain, Sriparna Saha, Pushpak
Bhattacharyya
- Abstract summary: Cyberbullying has become a big issue with the popularity of different social media networks and online communication apps.
Recent laws like "right to explanations" of General Data Protection Regulation have spurred research in developing interpretable models.
We develop first interpretable multi-task model called em mExCB for automatic cyberbullying detection from code-mixed languages.
- Score: 52.3781496277104
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Cyberbullying has become a big issue with the popularity of different social
media networks and online communication apps. While plenty of research is going
on to develop better models for cyberbullying detection in monolingual
language, there is very little research on the code-mixed languages and
explainability aspect of cyberbullying. Recent laws like "right to
explanations" of General Data Protection Regulation, have spurred research in
developing interpretable models rather than focusing on performance. Motivated
by this we develop the first interpretable multi-task model called {\em mExCB}
for automatic cyberbullying detection from code-mixed languages which can
simultaneously solve several tasks, cyberbullying detection,
explanation/rationale identification, target group detection and sentiment
analysis. We have introduced {\em BullyExplain}, the first benchmark dataset
for explainable cyberbullying detection in code-mixed language. Each post in
{\em BullyExplain} dataset is annotated with four labels, i.e., {\em bully
label, sentiment label, target and rationales (explainability)}, i.e., which
phrases are being responsible for annotating the post as a bully. The proposed
multitask framework (mExCB) based on CNN and GRU with word and sub-sentence
(SS) level attention is able to outperform several baselines and state of the
art models when applied on {\em BullyExplain} dataset.
Related papers
- Sentiment Analysis of Cyberbullying Data in Social Media [0.0]
Our work focuses on leveraging deep learning and natural language understanding techniques to detect traces of bullying in social media posts.
One approach utilizes BERT embeddings, while the other replaces the embeddings layer with the recently released embeddings API from OpenAI.
We conducted a performance comparison between these two approaches to evaluate their effectiveness in sentiment analysis of Formspring Cyberbullying data.
arXiv Detail & Related papers (2024-11-08T20:41:04Z) - The Use of a Large Language Model for Cyberbullying Detection [0.0]
cyberbullying (CB) is the most prevalent phenomenon in todays cyber world.
It is a severe threat to the mental and physical health of citizens.
This opens the need to develop a robust system to prevent bullying content from online forums, blogs, and social media platforms.
arXiv Detail & Related papers (2024-02-06T15:46:31Z) - Meme-ingful Analysis: Enhanced Understanding of Cyberbullying in Memes
Through Multimodal Explanations [48.82168723932981]
We introduce em MultiBully-Ex, the first benchmark dataset for multimodal explanation from code-mixed cyberbullying memes.
A Contrastive Language-Image Pretraining (CLIP) approach has been proposed for visual and textual explanation of a meme.
arXiv Detail & Related papers (2024-01-18T11:24:30Z) - How to Solve Few-Shot Abusive Content Detection Using the Data We Actually Have [58.23138483086277]
In this work we leverage datasets we already have, covering a wide range of tasks related to abusive language detection.
Our goal is to build models cheaply for a new target label set and/or language, using only a few training examples of the target domain.
Our experiments show that using already existing datasets and only a few-shots of the target task the performance of models improve both monolingually and across languages.
arXiv Detail & Related papers (2023-05-23T14:04:12Z) - Verifying the Robustness of Automatic Credibility Assessment [50.55687778699995]
We show that meaning-preserving changes in input text can mislead the models.
We also introduce BODEGA: a benchmark for testing both victim models and attack methods on misinformation detection tasks.
Our experimental results show that modern large language models are often more vulnerable to attacks than previous, smaller solutions.
arXiv Detail & Related papers (2023-03-14T16:11:47Z) - Addressing the Challenges of Cross-Lingual Hate Speech Detection [115.1352779982269]
In this paper we focus on cross-lingual transfer learning to support hate speech detection in low-resource languages.
We leverage cross-lingual word embeddings to train our neural network systems on the source language and apply it to the target language.
We investigate the issue of label imbalance of hate speech datasets, since the high ratio of non-hate examples compared to hate examples often leads to low model performance.
arXiv Detail & Related papers (2022-01-15T20:48:14Z) - Pseudo Siamese Network for Few-shot Intent Generation [54.10596778418007]
We propose a Pseudo Siamese Network (PSN) to generate labeled data for few-shot intents.
PSN consists of two identicalworks with the same structure but different weights: an action network and an object network.
Experiments on two real-world datasets show that PSN achieves state-of-the-art performance for the generalized few shot intent detection task.
arXiv Detail & Related papers (2021-05-03T14:30:47Z) - Analysing Cyberbullying using Natural Language Processing by
Understanding Jargon in Social Media [4.932130498861987]
In our work, we explore binary classification by using a combination of datasets from various social media platforms.
We experiment through multiple models such as Bi-LSTM, GloVe, state-of-the-art models like BERT, and apply a unique preprocessing technique by introducing a slang-abusive corpus.
arXiv Detail & Related papers (2021-04-23T04:20:19Z) - HENIN: Learning Heterogeneous Neural Interaction Networks for
Explainable Cyberbullying Detection on Social Media [11.443698975923176]
We propose a novel deep model, HEterogeneous Neural Interaction Networks (HENIN), for explainable cyberbullying detection.
HENIN contains the following components: a comment encoder, a post-comment co-attention sub-network, and session-session and post-post interaction extractors.
arXiv Detail & Related papers (2020-10-09T13:44:34Z) - Aggressive, Repetitive, Intentional, Visible, and Imbalanced: Refining
Representations for Cyberbullying Classification [4.945634077636197]
We study the nuanced problem of cyberbullying using five explicit factors to represent its social and linguistic aspects.
These results demonstrate the importance of representing and modeling cyberbullying as a social phenomenon.
arXiv Detail & Related papers (2020-04-04T00:35:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.