Detecting Troll Behavior via Inverse Reinforcement Learning: A Case
Study of Russian Trolls in the 2016 US Election
- URL: http://arxiv.org/abs/2001.10570v3
- Date: Fri, 5 Jun 2020 17:43:28 GMT
- Title: Detecting Troll Behavior via Inverse Reinforcement Learning: A Case
Study of Russian Trolls in the 2016 US Election
- Authors: Luca Luceri, Silvia Giordano, Emilio Ferrara
- Abstract summary: We propose an approach based on Inverse Reinforcement Learning (IRL) to capture troll behavior and identify troll accounts.
As a study case, we consider the troll accounts identified by the US Congress during the investigation of Russian meddling in the 2016 US Presidential election.
We report promising results: the IRL-based approach is able to accurately detect troll accounts.
- Score: 8.332032237125897
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Since the 2016 US Presidential election, social media abuse has been
eliciting massive concern in the academic community and beyond. Preventing and
limiting the malicious activity of users, such as trolls and bots, in their
manipulation campaigns is of paramount importance for the integrity of
democracy, public health, and more. However, the automated detection of troll
accounts is an open challenge. In this work, we propose an approach based on
Inverse Reinforcement Learning (IRL) to capture troll behavior and identify
troll accounts. We employ IRL to infer a set of online incentives that may
steer user behavior, which in turn highlights behavioral differences between
troll and non-troll accounts, enabling their accurate classification. As a
study case, we consider the troll accounts identified by the US Congress during
the investigation of Russian meddling in the 2016 US Presidential election. We
report promising results: the IRL-based approach is able to accurately detect
troll accounts (AUC=89.1%). The differences in the predictive features between
the two classes of accounts enables a principled understanding of the
distinctive behaviors reflecting the incentives trolls and non-trolls respond
to.
Related papers
- Coordinated Reply Attacks in Influence Operations: Characterization and Detection [43.98568073610101]
We characterize coordinated reply attacks in the context of influence operations on Twitter.
Our analysis reveals that the primary targets of these attacks are influential people such as journalists, news media, state officials, and politicians.
We propose two supervised machine-learning models, one to classify tweets to determine whether they are targeted by a reply attack, and one to classify accounts that reply to a targeted tweet to determine whether they are part of a coordinated attack.
arXiv Detail & Related papers (2024-10-25T02:57:08Z) - Towards Effective Counter-Responses: Aligning Human Preferences with Strategies to Combat Online Trolling [9.598920004159696]
This paper investigates whether humans have preferred strategies tailored to different types of trolling behaviors.
We introduce a methodology for generating counter-responses to trolls by recommending appropriate RSs.
The experimental results demonstrate that our proposed approach guides constructive discussion and reduces the negative effects of trolls.
arXiv Detail & Related papers (2024-10-05T14:01:52Z) - Russo-Ukrainian War: Prediction and explanation of Twitter suspension [47.61306219245444]
This study focuses on the Twitter suspension mechanism and the analysis of shared content and features of user accounts that may lead to this.
We have obtained a dataset containing 107.7M tweets, originating from 9.8 million users, using Twitter API.
Our results reveal scam campaigns taking advantage of trending topics regarding the Russia-Ukrainian conflict for Bitcoin fraud, spam, and advertisement campaigns.
arXiv Detail & Related papers (2023-06-06T08:41:02Z) - Characterizing, Detecting, and Predicting Online Ban Evasion [9.949354222717773]
Malicious users can easily create a new account to evade online bans.
We conduct the first data-driven study of ban evasion, i.e., the act of circumventing bans on an online platform.
We find that evasion child accounts demonstrate similarities with respect to their banned parent accounts on several behavioral axes.
arXiv Detail & Related papers (2022-02-10T18:58:19Z) - Identification of Twitter Bots based on an Explainable ML Framework: the
US 2020 Elections Case Study [72.61531092316092]
This paper focuses on the design of a novel system for identifying Twitter bots based on labeled Twitter data.
Supervised machine learning (ML) framework is adopted using an Extreme Gradient Boosting (XGBoost) algorithm.
Our study also deploys Shapley Additive Explanations (SHAP) for explaining the ML model predictions.
arXiv Detail & Related papers (2021-12-08T14:12:24Z) - TROLLMAGNIFIER: Detecting State-Sponsored Troll Accounts on Reddit [11.319938541673578]
We present TROLLMAGNIFIER, a detection system for troll accounts.
TROLLMAGNIFIER learns the typical behavior of known troll accounts and identifies more that behave similarly.
We show that using TROLLMAGNIFIER, one can grow the initial knowledge of potential trolls by over 300%.
arXiv Detail & Related papers (2021-12-01T12:10:24Z) - Exposing Paid Opinion Manipulation Trolls [19.834000431578737]
We show how to find paid trolls on the Web using machine learning.
In this paper, we assume that a user who is called a troll by several different people is likely to be such.
We compare the profiles of paid trolls vs. (ii)"mentioned" trolls vs. (iii) non-trolls, and we further show that a classifier trained to distinguish (ii) from (iii) does quite well also at telling apart (i) from (iii)
arXiv Detail & Related papers (2021-09-26T11:40:14Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - TrollHunter2020: Real-Time Detection of Trolling Narratives on Twitter
During the 2020 US Elections [1.5469452301122175]
TrollHunter2020 is a real-time detection mechanism used to hunt for trolling narratives on Twitter during the 2020 U.S. elections.
Our results suggest that the TrollHunter 2020 indeed captures the emerging trolling narratives in a very early stage of an unfolding polarizing event.
arXiv Detail & Related papers (2020-12-04T14:03:06Z) - Causal Understanding of Fake News Dissemination on Social Media [50.4854427067898]
We argue that it is critical to understand what user attributes potentially cause users to share fake news.
In fake news dissemination, confounders can be characterized by fake news sharing behavior that inherently relates to user attributes and online activities.
We propose a principled approach to alleviating selection bias in fake news dissemination.
arXiv Detail & Related papers (2020-10-20T19:37:04Z) - Russian trolls speaking Russian: Regional Twitter operations and MH17 [68.8204255655161]
In 2018, Twitter released data on accounts identified as Russian trolls.
We analyze the Russian-language operations of these trolls.
We find that trolls' information campaign on the MH17 crash was the largest in terms of tweet count.
arXiv Detail & Related papers (2020-05-13T19:48:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.