Viable Threat on News Reading: Generating Biased News Using Natural
Language Models
- URL: http://arxiv.org/abs/2010.02150v1
- Date: Mon, 5 Oct 2020 16:55:39 GMT
- Title: Viable Threat on News Reading: Generating Biased News Using Natural
Language Models
- Authors: Saurabh Gupta, Huy H. Nguyen, Junichi Yamagishi and Isao Echizen
- Abstract summary: We show that publicly available language models can reliably generate biased news content based on an input original news.
We also show that a large number of high-quality biased news articles can be generated using controllable text generation.
- Score: 49.90665530780664
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advancements in natural language generation has raised serious
concerns. High-performance language models are widely used for language
generation tasks because they are able to produce fluent and meaningful
sentences. These models are already being used to create fake news. They can
also be exploited to generate biased news, which can then be used to attack
news aggregators to change their reader's behavior and influence their bias. In
this paper, we use a threat model to demonstrate that the publicly available
language models can reliably generate biased news content based on an input
original news. We also show that a large number of high-quality biased news
articles can be generated using controllable text generation. A subjective
evaluation with 80 participants demonstrated that the generated biased news is
generally fluent, and a bias evaluation with 24 participants demonstrated that
the bias (left or right) is usually evident in the generated articles and can
be easily identified.
Related papers
- DocNet: Semantic Structure in Inductive Bias Detection Models [0.4779196219827508]
In this paper, we explore an often overlooked aspect of bias detection in documents: the semantic structure of news articles.
We present DocNet, a novel, inductive, and low-resource document embedding and bias detection model.
We also demonstrate that the semantic structure of news articles from opposing partisan sides, as represented in document-level graph embeddings, have significant similarities.
arXiv Detail & Related papers (2024-06-16T14:51:12Z) - Adapting Fake News Detection to the Era of Large Language Models [48.5847914481222]
We study the interplay between machine-(paraphrased) real news, machine-generated fake news, human-written fake news, and human-written real news.
Our experiments reveal an interesting pattern that detectors trained exclusively on human-written articles can indeed perform well at detecting machine-generated fake news, but not vice versa.
arXiv Detail & Related papers (2023-11-02T08:39:45Z) - Multilingual Coarse Political Stance Classification of Media. The
Editorial Line of a ChatGPT and Bard Newspaper [1.450405446885067]
We use ratings of authentic news outlets to create a multilingual corpus of news with coarse stance annotations.
We show that classifiers trained on this data are able to identify the editorial line of most unseen newspapers in English, German, Spanish and Catalan.
We observe that, similarly to traditional newspapers, ChatGPT editorial line evolves with time and, being a data-driven system, the stance of the generated articles differs among languages.
arXiv Detail & Related papers (2023-10-25T01:01:28Z) - Towards Corpus-Scale Discovery of Selection Biases in News Coverage:
Comparing What Sources Say About Entities as a Start [65.28355014154549]
This paper investigates the challenges of building scalable NLP systems for discovering patterns of media selection biases directly from news content in massive-scale news corpora.
We show the capabilities of the framework through a case study on NELA-2020, a corpus of 1.8M news articles in English from 519 news sources worldwide.
arXiv Detail & Related papers (2023-04-06T23:36:45Z) - Unveiling the Hidden Agenda: Biases in News Reporting and Consumption [59.55900146668931]
We build a six-year dataset on the Italian vaccine debate and adopt a Bayesian latent space model to identify narrative and selection biases.
We found a nonlinear relationship between biases and engagement, with higher engagement for extreme positions.
Analysis of news consumption on Twitter reveals common audiences among news outlets with similar ideological positions.
arXiv Detail & Related papers (2023-01-14T18:58:42Z) - Multiverse: Multilingual Evidence for Fake News Detection [71.51905606492376]
Multiverse is a new feature based on multilingual evidence that can be used for fake news detection.
The hypothesis of the usage of cross-lingual evidence as a feature for fake news detection is confirmed.
arXiv Detail & Related papers (2022-11-25T18:24:17Z) - Faking Fake News for Real Fake News Detection: Propaganda-loaded
Training Data Generation [105.20743048379387]
We propose a novel framework for generating training examples informed by the known styles and strategies of human-authored propaganda.
Specifically, we perform self-critical sequence training guided by natural language inference to ensure the validity of the generated articles.
Our experimental results show that fake news detectors trained on PropaNews are better at detecting human-written disinformation by 3.62 - 7.69% F1 score on two public datasets.
arXiv Detail & Related papers (2022-03-10T14:24:19Z) - Newsalyze: Enabling News Consumers to Understand Media Bias [7.652448987187803]
Knowing a news article's slant and authenticity is of crucial importance in times of "fake news"
We introduce Newsalyze, a bias-aware news reader focusing on a subtle, yet powerful form of media bias, named bias by word choice and labeling (WCL)
WCL bias can alter the assessment of entities reported in the news, e.g., "freedom fighters" vs. "terrorists"
arXiv Detail & Related papers (2021-05-20T11:20:37Z) - Analyzing Political Bias and Unfairness in News Articles at Different
Levels of Granularity [35.19976910093135]
The research presented in this paper addresses not only the automatic detection of bias but goes one step further in that it explores how political bias and unfairness are manifested linguistically.
We utilize a new corpus of 6964 news articles with labels derived from adfontesmedia.com and develop a neural model for bias assessment.
arXiv Detail & Related papers (2020-10-20T22:25:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.