AbuseAnalyzer: Abuse Detection, Severity and Target Prediction for Gab
Posts
- URL: http://arxiv.org/abs/2010.00038v2
- Date: Thu, 8 Oct 2020 17:42:33 GMT
- Title: AbuseAnalyzer: Abuse Detection, Severity and Target Prediction for Gab
Posts
- Authors: Mohit Chandra, Ashwin Pathak, Eesha Dutta, Paryul Jain, Manish Gupta,
Manish Shrivastava, Ponnurangam Kumaraguru
- Abstract summary: We present a first of the kind dataset with 7601 posts from Gab which looks at online abuse from the perspective of presence of abuse, severity and target of abusive behavior.
We also propose a system to address these tasks, obtaining an accuracy of 80% for abuse presence, 82% for abuse target prediction, and 65% for abuse severity prediction.
- Score: 19.32095911241636
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: While extensive popularity of online social media platforms has made
information dissemination faster, it has also resulted in widespread online
abuse of different types like hate speech, offensive language, sexist and
racist opinions, etc. Detection and curtailment of such abusive content is
critical for avoiding its psychological impact on victim communities, and
thereby preventing hate crimes. Previous works have focused on classifying user
posts into various forms of abusive behavior. But there has hardly been any
focus on estimating the severity of abuse and the target. In this paper, we
present a first of the kind dataset with 7601 posts from Gab which looks at
online abuse from the perspective of presence of abuse, severity and target of
abusive behavior. We also propose a system to address these tasks, obtaining an
accuracy of ~80% for abuse presence, ~82% for abuse target prediction, and ~65%
for abuse severity prediction.
Related papers
- MisinfoEval: Generative AI in the Era of "Alternative Facts" [50.069577397751175]
We introduce a framework for generating and evaluating large language model (LLM) based misinformation interventions.
We present (1) an experiment with a simulated social media environment to measure effectiveness of misinformation interventions, and (2) a second experiment with personalized explanations tailored to the demographics and beliefs of users.
Our findings confirm that LLM-based interventions are highly effective at correcting user behavior.
arXiv Detail & Related papers (2024-10-13T18:16:50Z) - Journalists are most likely to receive abuse: Analysing online abuse of UK public figures across sport, politics, and journalism on Twitter [0.0]
We present analysis of a novel dataset of 45.5M tweets targeted at 4,602 UK public figures across 3 domains.
We show that MPs receive more abuse in absolute terms, but that journalists are most likely to receive abuse after controlling for other factors.
We also find that a more prominent online presence and being male are indicative of higher levels of abuse across all 3 domains.
arXiv Detail & Related papers (2024-09-05T09:27:56Z) - Exploratory Data Analysis on Code-mixed Misogynistic Comments [0.0]
We present a novel dataset of YouTube comments in mix-code Hinglish.
These comments have been weak labelled as Misogynistic' and Non-misogynistic'
arXiv Detail & Related papers (2024-03-09T23:21:17Z) - Analyzing Norm Violations in Live-Stream Chat [49.120561596550395]
We study the first NLP study dedicated to detecting norm violations in conversations on live-streaming platforms.
We define norm violation categories in live-stream chats and annotate 4,583 moderated comments from Twitch.
Our results show that appropriate contextual information can boost moderation performance by 35%.
arXiv Detail & Related papers (2023-05-18T05:58:27Z) - Explainable Abuse Detection as Intent Classification and Slot Filling [66.80201541759409]
We introduce the concept of policy-aware abuse detection, abandoning the unrealistic expectation that systems can reliably learn which phenomena constitute abuse from inspecting the data alone.
We show how architectures for intent classification and slot filling can be used for abuse detection, while providing a rationale for model decisions.
arXiv Detail & Related papers (2022-10-06T03:33:30Z) - Enriching Abusive Language Detection with Community Context [0.3708656266586145]
Use of pejorative expressions can be benign or actively empowering.
Models for abuse detection misclassify these expressions as derogatory, inadvertently censor productive conversations held by marginalized groups.
Our paper highlights how community context can improve classification outcomes in abusive language detection.
arXiv Detail & Related papers (2022-06-16T20:54:02Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - The User behind the Abuse: A Position on Ethics and Explainability [25.791014642037585]
We discuss the role that modeling of users and online communities plays in abuse detection.
We then explore the ethical challenges of incorporating user and community information.
We propose properties that an explainable method should aim to exhibit.
arXiv Detail & Related papers (2021-03-31T16:20:37Z) - Detecting Harmful Content On Online Platforms: What Platforms Need Vs.
Where Research Efforts Go [44.774035806004214]
harmful content on online platforms comes in many different forms including hate speech, offensive language, bullying and harassment, misinformation, spam, violence, graphic content, sexual abuse, self harm, and many other.
Online platforms seek to moderate such content to limit societal harm, to comply with legislation, and to create a more inclusive environment for their users.
There is currently a dichotomy between what types of harmful content online platforms seek to curb, and what research efforts there are to automatically detect such content.
arXiv Detail & Related papers (2021-02-27T08:01:10Z) - ETHOS: an Online Hate Speech Detection Dataset [6.59720246184989]
We present 'ETHOS', a textual dataset with two variants: binary and multi-label, based on YouTube and Reddit comments validated using the Figure-Eight crowdsourcing platform.
Our key assumption is that, even gaining a small amount of labelled data from such a time-consuming process, we can guarantee hate speech occurrences in the examined material.
arXiv Detail & Related papers (2020-06-11T08:59:57Z) - Information Consumption and Social Response in a Segregated Environment:
the Case of Gab [74.5095691235917]
This work provides a characterization of the interaction patterns within Gab around the COVID-19 topic.
We find that there are no strong statistical differences in the social response to questionable and reliable content.
Our results provide insights toward the understanding of coordinated inauthentic behavior and on the early-warning of information operation.
arXiv Detail & Related papers (2020-06-03T11:34:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.