Detecting Troll Behavior via Inverse Reinforcement Learning: A Case
Study of Russian Trolls in the 2016 US Election
- URL: http://arxiv.org/abs/2001.10570v3
- Date: Fri, 5 Jun 2020 17:43:28 GMT
- Title: Detecting Troll Behavior via Inverse Reinforcement Learning: A Case
Study of Russian Trolls in the 2016 US Election
- Authors: Luca Luceri, Silvia Giordano, Emilio Ferrara
- Abstract summary: We propose an approach based on Inverse Reinforcement Learning (IRL) to capture troll behavior and identify troll accounts.
As a study case, we consider the troll accounts identified by the US Congress during the investigation of Russian meddling in the 2016 US Presidential election.
We report promising results: the IRL-based approach is able to accurately detect troll accounts.
- Score: 8.332032237125897
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Since the 2016 US Presidential election, social media abuse has been
eliciting massive concern in the academic community and beyond. Preventing and
limiting the malicious activity of users, such as trolls and bots, in their
manipulation campaigns is of paramount importance for the integrity of
democracy, public health, and more. However, the automated detection of troll
accounts is an open challenge. In this work, we propose an approach based on
Inverse Reinforcement Learning (IRL) to capture troll behavior and identify
troll accounts. We employ IRL to infer a set of online incentives that may
steer user behavior, which in turn highlights behavioral differences between
troll and non-troll accounts, enabling their accurate classification. As a
study case, we consider the troll accounts identified by the US Congress during
the investigation of Russian meddling in the 2016 US Presidential election. We
report promising results: the IRL-based approach is able to accurately detect
troll accounts (AUC=89.1%). The differences in the predictive features between
the two classes of accounts enables a principled understanding of the
distinctive behaviors reflecting the incentives trolls and non-trolls respond
to.
Related papers
- Russo-Ukrainian War: Prediction and explanation of Twitter suspension [47.61306219245444]
This study focuses on the Twitter suspension mechanism and the analysis of shared content and features of user accounts that may lead to this.
We have obtained a dataset containing 107.7M tweets, originating from 9.8 million users, using Twitter API.
Our results reveal scam campaigns taking advantage of trending topics regarding the Russia-Ukrainian conflict for Bitcoin fraud, spam, and advertisement campaigns.
arXiv Detail & Related papers (2023-06-06T08:41:02Z) - Exposing Influence Campaigns in the Age of LLMs: A Behavioral-Based AI
Approach to Detecting State-Sponsored Trolls [8.202465737306222]
Detection of state-sponsored trolls operating in influence campaigns on social media is a critical and unsolved challenge.
We propose a new AI-based solution that identifies troll accounts solely through behavioral cues associated with their sequences of sharing activity.
arXiv Detail & Related papers (2022-10-17T07:01:17Z) - Illusory Attacks: Information-Theoretic Detectability Matters in Adversarial Attacks [76.35478518372692]
We introduce epsilon-illusory, a novel form of adversarial attack on sequential decision-makers.
Compared to existing attacks, we empirically find epsilon-illusory to be significantly harder to detect with automated methods.
Our findings suggest the need for better anomaly detectors, as well as effective hardware- and system-level defenses.
arXiv Detail & Related papers (2022-07-20T19:49:09Z) - Characterizing, Detecting, and Predicting Online Ban Evasion [9.949354222717773]
Malicious users can easily create a new account to evade online bans.
We conduct the first data-driven study of ban evasion, i.e., the act of circumventing bans on an online platform.
We find that evasion child accounts demonstrate similarities with respect to their banned parent accounts on several behavioral axes.
arXiv Detail & Related papers (2022-02-10T18:58:19Z) - Identification of Twitter Bots based on an Explainable ML Framework: the
US 2020 Elections Case Study [72.61531092316092]
This paper focuses on the design of a novel system for identifying Twitter bots based on labeled Twitter data.
Supervised machine learning (ML) framework is adopted using an Extreme Gradient Boosting (XGBoost) algorithm.
Our study also deploys Shapley Additive Explanations (SHAP) for explaining the ML model predictions.
arXiv Detail & Related papers (2021-12-08T14:12:24Z) - TROLLMAGNIFIER: Detecting State-Sponsored Troll Accounts on Reddit [11.319938541673578]
We present TROLLMAGNIFIER, a detection system for troll accounts.
TROLLMAGNIFIER learns the typical behavior of known troll accounts and identifies more that behave similarly.
We show that using TROLLMAGNIFIER, one can grow the initial knowledge of potential trolls by over 300%.
arXiv Detail & Related papers (2021-12-01T12:10:24Z) - Exposing Paid Opinion Manipulation Trolls [19.834000431578737]
We show how to find paid trolls on the Web using machine learning.
In this paper, we assume that a user who is called a troll by several different people is likely to be such.
We compare the profiles of paid trolls vs. (ii)"mentioned" trolls vs. (iii) non-trolls, and we further show that a classifier trained to distinguish (ii) from (iii) does quite well also at telling apart (i) from (iii)
arXiv Detail & Related papers (2021-09-26T11:40:14Z) - News consumption and social media regulations policy [70.31753171707005]
We analyze two social media that enforced opposite moderation methods, Twitter and Gab, to assess the interplay between news consumption and content regulation.
Our results show that the presence of moderation pursued by Twitter produces a significant reduction of questionable content.
The lack of clear regulation on Gab results in the tendency of the user to engage with both types of content, showing a slight preference for the questionable ones which may account for a dissing/endorsement behavior.
arXiv Detail & Related papers (2021-06-07T19:26:32Z) - TrollHunter2020: Real-Time Detection of Trolling Narratives on Twitter
During the 2020 US Elections [1.5469452301122175]
TrollHunter2020 is a real-time detection mechanism used to hunt for trolling narratives on Twitter during the 2020 U.S. elections.
Our results suggest that the TrollHunter 2020 indeed captures the emerging trolling narratives in a very early stage of an unfolding polarizing event.
arXiv Detail & Related papers (2020-12-04T14:03:06Z) - Causal Understanding of Fake News Dissemination on Social Media [50.4854427067898]
We argue that it is critical to understand what user attributes potentially cause users to share fake news.
In fake news dissemination, confounders can be characterized by fake news sharing behavior that inherently relates to user attributes and online activities.
We propose a principled approach to alleviating selection bias in fake news dissemination.
arXiv Detail & Related papers (2020-10-20T19:37:04Z) - Russian trolls speaking Russian: Regional Twitter operations and MH17 [68.8204255655161]
In 2018, Twitter released data on accounts identified as Russian trolls.
We analyze the Russian-language operations of these trolls.
We find that trolls' information campaign on the MH17 crash was the largest in terms of tweet count.
arXiv Detail & Related papers (2020-05-13T19:48:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.