Detecting Online Hate Speech: Approaches Using Weak Supervision and
Network Embedding Models
- URL: http://arxiv.org/abs/2007.12724v1
- Date: Fri, 24 Jul 2020 18:13:52 GMT
- Title: Detecting Online Hate Speech: Approaches Using Weak Supervision and
Network Embedding Models
- Authors: Michael Ridenhour, Arunkumar Bagavathi, Elaheh Raisi, Siddharth
Krishnan
- Abstract summary: We propose a weak supervision deep learning model that quantitatively uncover hateful users and (ii) present a novel qualitative analysis to uncover indirect hateful conversations.
We evaluate our model on 19.2M posts and show that our weak supervision model outperforms the baseline models in identifying indirect hateful interactions.
We also analyze a multilayer network, constructed from two types of user interactions in Gab(quote and reply) and interaction scores from the weak supervision model as edge weights, to predict hateful users.
- Score: 2.3322477552758234
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The ubiquity of social media has transformed online interactions among
individuals. Despite positive effects, it has also allowed anti-social elements
to unite in alternative social media environments (eg. Gab.com) like never
before. Detecting such hateful speech using automated techniques can allow
social media platforms to moderate their content and prevent nefarious
activities like hate speech propagation. In this work, we propose a weak
supervision deep learning model that - (i) quantitatively uncover hateful users
and (ii) present a novel qualitative analysis to uncover indirect hateful
conversations. This model scores content on the interaction level, rather than
the post or user level, and allows for characterization of users who most
frequently participate in hateful conversations. We evaluate our model on 19.2M
posts and show that our weak supervision model outperforms the baseline models
in identifying indirect hateful interactions. We also analyze a multilayer
network, constructed from two types of user interactions in Gab(quote and
reply) and interaction scores from the weak supervision model as edge weights,
to predict hateful users. We utilize the multilayer network embedding methods
to generate features for the prediction task and we show that considering user
context from multiple networks help achieving better predictions of hateful
users in Gab. We receive up to 7% performance gain compared to single layer or
homogeneous network embedding models.
Related papers
- AggregHate: An Efficient Aggregative Approach for the Detection of Hatemongers on Social Platforms [4.649475179575046]
We consider a multimodal aggregative approach for the detection of hate-mongers, taking into account the potentially hateful texts, user activity, and the user network.
Our method can be used to improve the classification of coded messages, dog-whistling, and racial gas-lighting, as well as inform intervention measures.
arXiv Detail & Related papers (2024-09-22T14:29:49Z) - Decoding the Silent Majority: Inducing Belief Augmented Social Graph
with Large Language Model for Response Forecasting [74.68371461260946]
SocialSense is a framework that induces a belief-centered graph on top of an existent social network, along with graph-based propagation to capture social dynamics.
Our method surpasses existing state-of-the-art in experimental evaluations for both zero-shot and supervised settings.
arXiv Detail & Related papers (2023-10-20T06:17:02Z) - Understanding writing style in social media with a supervised
contrastively pre-trained transformer [57.48690310135374]
Online Social Networks serve as fertile ground for harmful behavior, ranging from hate speech to the dissemination of disinformation.
We introduce the Style Transformer for Authorship Representations (STAR), trained on a large corpus derived from public sources of 4.5 x 106 authored texts.
Using a support base of 8 documents of 512 tokens, we can discern authors from sets of up to 1616 authors with at least 80% accuracy.
arXiv Detail & Related papers (2023-10-17T09:01:17Z) - CoSyn: Detecting Implicit Hate Speech in Online Conversations Using a
Context Synergized Hyperbolic Network [52.85130555886915]
CoSyn is a context-synergized neural network that explicitly incorporates user- and conversational context for detecting implicit hate speech in online conversations.
We show that CoSyn outperforms all our baselines in detecting implicit hate speech with absolute improvements in the range of 1.24% - 57.8%.
arXiv Detail & Related papers (2023-03-02T17:30:43Z) - Countering Malicious Content Moderation Evasion in Online Social
Networks: Simulation and Detection of Word Camouflage [64.78260098263489]
Twisting and camouflaging keywords are among the most used techniques to evade platform content moderation systems.
This article contributes significantly to countering malicious information by developing multilingual tools to simulate and detect new methods of evasion of content.
arXiv Detail & Related papers (2022-12-27T16:08:49Z) - CRUSH: Contextually Regularized and User anchored Self-supervised Hate
speech Detection [6.759148939470331]
We introduce CRUSH, a framework for hate speech detection using user-anchored self-supervision and contextual regularization.
Our proposed approach secures 1-12% improvement in test set metrics over best performing previous approaches on two types of tasks and multiple popular english social media datasets.
arXiv Detail & Related papers (2022-04-13T13:51:51Z) - Addressing the Challenges of Cross-Lingual Hate Speech Detection [115.1352779982269]
In this paper we focus on cross-lingual transfer learning to support hate speech detection in low-resource languages.
We leverage cross-lingual word embeddings to train our neural network systems on the source language and apply it to the target language.
We investigate the issue of label imbalance of hate speech datasets, since the high ratio of non-hate examples compared to hate examples often leads to low model performance.
arXiv Detail & Related papers (2022-01-15T20:48:14Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - Leveraging cross-platform data to improve automated hate speech
detection [0.0]
Most existing approaches for hate speech detection focus on a single social media platform in isolation.
Here we propose a new cross-platform approach to detect hate speech which leverages multiple datasets and classification models from different platforms.
We demonstrate how this approach outperforms existing models, and achieves good performance when tested on messages from novel social media platforms.
arXiv Detail & Related papers (2021-02-09T15:52:34Z) - An Online Multilingual Hate speech Recognition System [13.87667165678441]
We analyse six datasets by combining them into a single homogeneous dataset and classify them into three classes, abusive, hateful or neither.
We create a tool which identifies and scores a page with effective metric in near-real time and uses the same as feedback to re-train our model.
We prove the competitive performance of our multilingual model on two langauges, English and Hindi, leading to comparable or superior performance to most monolingual models.
arXiv Detail & Related papers (2020-11-23T16:33:48Z) - Transfer Learning for Hate Speech Detection in Social Media [14.759208309842178]
This paper uses a transfer learning technique to leverage two independent datasets jointly.
We build an interpretable two-dimensional visualization tool of the constructed hate speech representation -- dubbed the Map of Hate.
We show that the joint representation boosts prediction performances when only a limited amount of supervision is available.
arXiv Detail & Related papers (2019-06-10T08:00:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.