Fake News Spreader Detection on Twitter using Character N-Grams.
Notebook for PAN at CLEF 2020
- URL: http://arxiv.org/abs/2009.13859v1
- Date: Tue, 29 Sep 2020 08:32:32 GMT
- Title: Fake News Spreader Detection on Twitter using Character N-Grams.
Notebook for PAN at CLEF 2020
- Authors: Inna Vogel and Meghana Meghana
- Abstract summary: This notebook describes our profiling system for the fake news detection task on Twitter.
We conduct different feature extraction techniques and learning experiments from a multilingual perspective.
Our models achieve an overall accuracy of 73% and 79% on the English and Spanish official test set.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The authors of fake news often use facts from verified news sources and mix
them with misinformation to create confusion and provoke unrest among the
readers. The spread of fake news can thereby have serious implications on our
society. They can sway political elections, push down the stock price or crush
reputations of corporations or public figures. Several websites have taken on
the mission of checking rumors and allegations, but are often not fast enough
to check the content of all the news being disseminated. Especially social
media websites have offered an easy platform for the fast propagation of
information. Towards limiting fake news from being propagated among social
media users, the task of this year's PAN 2020 challenge lays the focus on the
fake news spreaders. The aim of the task is to determine whether it is possible
to discriminate authors that have shared fake news in the past from those that
have never done it. In this notebook, we describe our profiling system for the
fake news detection task on Twitter. For this, we conduct different feature
extraction techniques and learning experiments from a multilingual perspective,
namely English and Spanish. Our final submitted systems use character n-grams
as features in combination with a linear SVM for English and Logistic
Regression for the Spanish language. Our submitted models achieve an overall
accuracy of 73% and 79% on the English and Spanish official test set,
respectively. Our experiments show that it is difficult to differentiate
solidly fake news spreaders on Twitter from users who share credible
information leaving room for further investigations. Our model ranked 3rd out
of 72 competitors.
Related papers
- Adapting Fake News Detection to the Era of Large Language Models [48.5847914481222]
We study the interplay between machine-(paraphrased) real news, machine-generated fake news, human-written fake news, and human-written real news.
Our experiments reveal an interesting pattern that detectors trained exclusively on human-written articles can indeed perform well at detecting machine-generated fake news, but not vice versa.
arXiv Detail & Related papers (2023-11-02T08:39:45Z) - Fake News Detection and Behavioral Analysis: Case of COVID-19 [0.22940141855172028]
"Infodemic" due to spread of fake news regarding the pandemic has been a global issue.
Readers could mistake fake news for real news, and consequently have less access to authentic information.
It is challenging to accurately identify fake news data in social media posts.
arXiv Detail & Related papers (2023-05-25T13:42:08Z) - Multiverse: Multilingual Evidence for Fake News Detection [71.51905606492376]
Multiverse is a new feature based on multilingual evidence that can be used for fake news detection.
The hypothesis of the usage of cross-lingual evidence as a feature for fake news detection is confirmed.
arXiv Detail & Related papers (2022-11-25T18:24:17Z) - A Proposed Bi-LSTM Method to Fake News Detection [0.0]
False news was a determining factor in influencing the outcome of the U.S. presidential election.
Bi-LSTM was applied to determine if the news is false or real.
After creating & running the model, the work achieved 84% model accuracy and 62.0 F1-macro scores with training data.
arXiv Detail & Related papers (2022-06-15T06:36:42Z) - Faking Fake News for Real Fake News Detection: Propaganda-loaded
Training Data Generation [105.20743048379387]
We propose a novel framework for generating training examples informed by the known styles and strategies of human-authored propaganda.
Specifically, we perform self-critical sequence training guided by natural language inference to ensure the validity of the generated articles.
Our experimental results show that fake news detectors trained on PropaNews are better at detecting human-written disinformation by 3.62 - 7.69% F1 score on two public datasets.
arXiv Detail & Related papers (2022-03-10T14:24:19Z) - Fake News Detection Tools and Methods -- A Review [0.0]
We discuss the recent literature about different approaches to detect fake news over the Internet.
We highlight the various publicly available datasets and various online tools that are available and can debunk Fake News in real-time.
arXiv Detail & Related papers (2021-11-21T13:19:23Z) - Profiling Fake News Spreaders on Social Media through Psychological and
Motivational Factors [26.942545715296983]
We study the characteristics and motivational factors of fake news spreaders on social media.
We then perform a series of experiments to determine if fake news spreaders can be found to exhibit different characteristics than other users.
arXiv Detail & Related papers (2021-08-24T20:27:38Z) - Is it Fake? News Disinformation Detection on South African News Websites [0.015863809575305417]
Natural Language Processing is widely used in detecting fake news.
It is especially a problem in more localised contexts such as in South Africa.
In this work we investigate fake news detection on South African websites.
arXiv Detail & Related papers (2021-08-06T04:54:03Z) - User Preference-aware Fake News Detection [61.86175081368782]
Existing fake news detection algorithms focus on mining news content for deceptive signals.
We propose a new framework, UPFD, which simultaneously captures various signals from user preferences by joint content and graph modeling.
arXiv Detail & Related papers (2021-04-25T21:19:24Z) - The Rise and Fall of Fake News sites: A Traffic Analysis [62.51737815926007]
We investigate the online presence of fake news websites and characterize their behavior in comparison to real news websites.
Based on our findings, we build a content-agnostic ML for automatic detection of fake news websites.
arXiv Detail & Related papers (2021-03-16T18:10:22Z) - Causal Understanding of Fake News Dissemination on Social Media [50.4854427067898]
We argue that it is critical to understand what user attributes potentially cause users to share fake news.
In fake news dissemination, confounders can be characterized by fake news sharing behavior that inherently relates to user attributes and online activities.
We propose a principled approach to alleviating selection bias in fake news dissemination.
arXiv Detail & Related papers (2020-10-20T19:37:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.