Automatic Detection of Influential Actors in Disinformation Networks
- URL: http://arxiv.org/abs/2005.10879v3
- Date: Thu, 7 Jan 2021 22:15:57 GMT
- Title: Automatic Detection of Influential Actors in Disinformation Networks
- Authors: Steven T. Smith, Edward K. Kao, Erika D. Mackin, Danelle C. Shah, Olga
Simek, Donald B. Rubin
- Abstract summary: This paper presents an end-to-end framework to automate detection of disinformation narratives, networks, and influential actors.
System detects IO accounts with 96% precision, 79% recall, and 96% area-under-the-PR-curve.
Results are corroborated with independent sources of known IO accounts from U.S. Congressional reports, investigative journalism, and IO datasets provided by Twitter.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The weaponization of digital communications and social media to conduct
disinformation campaigns at immense scale, speed, and reach presents new
challenges to identify and counter hostile influence operations (IOs). This
paper presents an end-to-end framework to automate detection of disinformation
narratives, networks, and influential actors. The framework integrates natural
language processing, machine learning, graph analytics, and a novel network
causal inference approach to quantify the impact of individual actors in
spreading IO narratives. We demonstrate its capability on real-world hostile IO
campaigns with Twitter datasets collected during the 2017 French presidential
elections, and known IO accounts disclosed by Twitter over a broad range of IO
campaigns (May 2007 to February 2020), over 50,000 accounts, 17 countries, and
different account types including both trolls and bots. Our system detects IO
accounts with 96% precision, 79% recall, and 96% area-under-the-PR-curve, maps
out salient network communities, and discovers high-impact accounts that escape
the lens of traditional impact statistics based on activity counts and network
centrality. Results are corroborated with independent sources of known IO
accounts from U.S. Congressional reports, investigative journalism, and IO
datasets provided by Twitter.
Related papers
- Uncovering Coordinated Cross-Platform Information Operations Threatening the Integrity of the 2024 U.S. Presidential Election Online Discussion [8.557128766155229]
Information Operations pose a significant threat to the integrity of democratic processes.
In anticipation of the 2024 U.S. presidential election, we present a study aimed at uncovering the digital traces of coordinated IOs on $mathbbX$ (formerly Twitter)
Using our machine learning framework for detecting online coordination, we analyze a dataset comprising election-related conversations on $mathbbX$ from May 2024.
arXiv Detail & Related papers (2024-09-23T17:36:14Z) - Mapping the Russian Internet Troll Network on Twitter using a Predictive Model [0.0]
Russian Internet Trolls use fake personas to spread disinformation through multiple social media streams.
We create a predictive model to map the network operations.
Our model attains 88% prediction accuracy for the test set.
arXiv Detail & Related papers (2024-09-11T19:09:21Z) - Unraveling the Web of Disinformation: Exploring the Larger Context of State-Sponsored Influence Campaigns on Twitter [16.64763746842362]
We study 19 state-sponsored disinformation campaigns that took place on Twitter, originating from various countries.
We build a machine learning-based classifier that can correctly identify up to 94% of accounts from unseen campaigns.
We also run our system in the wild and find more accounts that could potentially belong to state-backed operations.
arXiv Detail & Related papers (2024-07-25T15:03:33Z) - Decoding the Silent Majority: Inducing Belief Augmented Social Graph
with Large Language Model for Response Forecasting [74.68371461260946]
SocialSense is a framework that induces a belief-centered graph on top of an existent social network, along with graph-based propagation to capture social dynamics.
Our method surpasses existing state-of-the-art in experimental evaluations for both zero-shot and supervised settings.
arXiv Detail & Related papers (2023-10-20T06:17:02Z) - ManiTweet: A New Benchmark for Identifying Manipulation of News on Social Media [74.93847489218008]
We present a novel task, identifying manipulation of news on social media, which aims to detect manipulation in social media posts and identify manipulated or inserted information.
To study this task, we have proposed a data collection schema and curated a dataset called ManiTweet, consisting of 3.6K pairs of tweets and corresponding articles.
Our analysis demonstrates that this task is highly challenging, with large language models (LLMs) yielding unsatisfactory performance.
arXiv Detail & Related papers (2023-05-23T16:40:07Z) - Design and analysis of tweet-based election models for the 2021 Mexican
legislative election [55.41644538483948]
We use a dataset of 15 million election-related tweets in the six months preceding election day.
We find that models using data with geographical attributes determine the results of the election with better precision and accuracy than conventional polling methods.
arXiv Detail & Related papers (2023-01-02T12:40:05Z) - Detecting Propagators of Disinformation on Twitter Using Quantitative
Discursive Analysis [0.0]
This study presents a method of identifying Russian disinformation bots on Twitter using resonance centering analysis and Clauset-Newman-Moore community detection.
The data reflect a significant degree of discursive dissimilarity between known Russian disinformation bots and a control set of Twitter users during the timeframe of the 2016 U.S. Presidential Election.
arXiv Detail & Related papers (2022-10-11T20:11:50Z) - Identification of Twitter Bots based on an Explainable ML Framework: the
US 2020 Elections Case Study [72.61531092316092]
This paper focuses on the design of a novel system for identifying Twitter bots based on labeled Twitter data.
Supervised machine learning (ML) framework is adopted using an Extreme Gradient Boosting (XGBoost) algorithm.
Our study also deploys Shapley Additive Explanations (SHAP) for explaining the ML model predictions.
arXiv Detail & Related papers (2021-12-08T14:12:24Z) - Shifting Polarization and Twitter News Influencers between two U.S.
Presidential Elections [92.33485580547801]
We analyze the change of polarization between the 2016 and 2020 U.S. presidential elections.
Most of the top influencers were affiliated with media organizations during both elections.
75% of the top influencers in 2020 were not present in 2016, demonstrating that such status is difficult to retain.
arXiv Detail & Related papers (2021-11-03T20:08:54Z) - FacTeR-Check: Semi-automated fact-checking through Semantic Similarity
and Natural Language Inference [61.068947982746224]
FacTeR-Check enables retrieving fact-checked information, unchecked claims verification and tracking dangerous information over social media.
The architecture is validated using a new dataset called NLI19-SP that is publicly released with COVID-19 related hoaxes and tweets from Spanish social media.
Our results show state-of-the-art performance on the individual benchmarks, as well as producing useful analysis of the evolution over time of 61 different hoaxes.
arXiv Detail & Related papers (2021-10-27T15:44:54Z) - SOK: Seeing and Believing: Evaluating the Trustworthiness of Twitter
Users [4.609388510200741]
Currently, there is no automated way of determining which news or users are credible and which are not.
In this work, we created a model which analysed the behaviour of50,000 politicians on Twitter.
We classified the political Twitter users as either trusted or untrusted using random forest, multilayer perceptron, and support vector machine.
arXiv Detail & Related papers (2021-07-16T17:39:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.