MediaSpin: Exploring Media Bias Through Fine-Grained Analysis of News Headlines
- URL: http://arxiv.org/abs/2412.02271v2
- Date: Fri, 23 May 2025 03:07:31 GMT
- Title: MediaSpin: Exploring Media Bias Through Fine-Grained Analysis of News Headlines
- Authors: Preetika Verma, Kokil Jaidka,
- Abstract summary: This study introduces the MediaSpin dataset, the first to characterize the bias in how prominent news outlets editorialize news headlines after publication.<n>We discuss the linguistic insights it affords and show its applications for bias prediction and user behavior analysis.
- Score: 12.636213065708318
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The editability of online news content has become a significant factor in shaping public perception, as social media platforms introduce new affordances for dynamic and adaptive news framing. Edits to news headlines can refocus audience attention, add or remove emotional language, and shift the framing of events in subtle yet impactful ways. What types of media bias are editorialized in and out of news headlines, and how can they be systematically identified? This study introduces the MediaSpin dataset, the first to characterize the bias in how prominent news outlets editorialize news headlines after publication. The dataset includes 78,910 pairs of headlines annotated with 13 distinct types of media bias, using human-supervised LLM labeling. We discuss the linguistic insights it affords and show its applications for bias prediction and user behavior analysis.
Related papers
- Unbiasing through Textual Descriptions: Mitigating Representation Bias in Video Benchmarks [85.54792243128695]
"Unbiased through Textual Description (UTD)" video benchmark based on unbiased subsets of existing video classification and retrieval datasets.
We leverage VLMs and LLMs to analyze and debias benchmarks from representation biases.
We conduct a systematic analysis of 12 popular video classification and retrieval datasets.
We benchmark 30 state-of-the-art video models on original and debiased splits and analyze biases in the models.
arXiv Detail & Related papers (2025-03-24T13:00:25Z) - Mapping the Media Landscape: Predicting Factual Reporting and Political Bias Through Web Interactions [0.7249731529275342]
We propose an extension to a recently presented news media reliability estimation method.
We assess the classification performance of four reinforcement learning strategies on a large news media hyperlink graph.
Our experiments, targeting two challenging bias descriptors, factual reporting and political bias, showed a significant performance improvement at the source media level.
arXiv Detail & Related papers (2024-10-23T08:18:26Z) - The Media Bias Taxonomy: A Systematic Literature Review on the Forms and
Automated Detection of Media Bias [5.579028648465784]
This article summarizes the research on computational methods to detect media bias by systematically reviewing 3140 research papers published between 2019 and 2022.
We show that media bias detection is a highly active research field, in which transformer-based classification approaches have led to significant improvements in recent years.
arXiv Detail & Related papers (2023-12-26T18:13:52Z) - Navigating News Narratives: A Media Bias Analysis Dataset [3.0821115746307672]
"Navigating News Narratives: A Media Bias Analysis dataset" is a comprehensive dataset to address the urgent need for tools to detect and analyze media bias.
This dataset encompasses a broad spectrum of biases, making it a unique and valuable asset in the field of media studies and artificial intelligence.
arXiv Detail & Related papers (2023-11-30T19:59:19Z) - Towards Corpus-Scale Discovery of Selection Biases in News Coverage:
Comparing What Sources Say About Entities as a Start [65.28355014154549]
This paper investigates the challenges of building scalable NLP systems for discovering patterns of media selection biases directly from news content in massive-scale news corpora.
We show the capabilities of the framework through a case study on NELA-2020, a corpus of 1.8M news articles in English from 519 news sources worldwide.
arXiv Detail & Related papers (2023-04-06T23:36:45Z) - Bias or Diversity? Unraveling Fine-Grained Thematic Discrepancy in U.S.
News Headlines [63.52264764099532]
We use a large dataset of 1.8 million news headlines from major U.S. media outlets spanning from 2014 to 2022.
We quantify the fine-grained thematic discrepancy related to four prominent topics - domestic politics, economic issues, social issues, and foreign affairs.
Our findings indicate that on domestic politics and social issues, the discrepancy can be attributed to a certain degree of media bias.
arXiv Detail & Related papers (2023-03-28T03:31:37Z) - Computational Assessment of Hyperpartisanship in News Titles [55.92100606666497]
We first adopt a human-guided machine learning framework to develop a new dataset for hyperpartisan news title detection.
Overall the Right media tends to use proportionally more hyperpartisan titles.
We identify three major topics including foreign issues, political systems, and societal issues that are suggestive of hyperpartisanship in news titles.
arXiv Detail & Related papers (2023-01-16T05:56:58Z) - Unveiling the Hidden Agenda: Biases in News Reporting and Consumption [59.55900146668931]
We build a six-year dataset on the Italian vaccine debate and adopt a Bayesian latent space model to identify narrative and selection biases.
We found a nonlinear relationship between biases and engagement, with higher engagement for extreme positions.
Analysis of news consumption on Twitter reveals common audiences among news outlets with similar ideological positions.
arXiv Detail & Related papers (2023-01-14T18:58:42Z) - Neural Media Bias Detection Using Distant Supervision With BABE -- Bias
Annotations By Experts [24.51774048437496]
This paper presents BABE, a robust and diverse data set for media bias research.
It consists of 3,700 sentences balanced among topics and outlets, containing media bias labels on the word and sentence level.
Based on our data, we also introduce a way to detect bias-inducing sentences in news articles automatically.
arXiv Detail & Related papers (2022-09-29T05:32:55Z) - NeuS: Neutral Multi-News Summarization for Mitigating Framing Bias [54.89737992911079]
We propose a new task, a neutral summary generation from multiple news headlines of the varying political spectrum.
One of the most interesting observations is that generation models can hallucinate not only factually inaccurate or unverifiable content, but also politically biased content.
arXiv Detail & Related papers (2022-04-11T07:06:01Z) - Newsalyze: Effective Communication of Person-Targeting Biases in News
Articles [8.586057042714698]
We present a system for bias identification, which combines state-of-the-art methods from natural language understanding.
Second, we devise bias-sensitive visualizations to communicate bias in news articles to non-expert news consumers.
Third, our main contribution is a large-scale user study that measures bias-awareness in a setting that approximates daily news consumption.
arXiv Detail & Related papers (2021-10-18T10:23:19Z) - MBIC -- A Media Bias Annotation Dataset Including Annotator
Characteristics [0.0]
Media bias, or slanted news coverage, can have a substantial impact on public perception of events.
In this poster, we present a matrix-based methodology to crowdsource such data using a self-developed annotation platform.
We also present MBIC - the first sample of 1,700 statements representing various media bias instances.
arXiv Detail & Related papers (2021-05-20T15:05:17Z) - Newsalyze: Enabling News Consumers to Understand Media Bias [7.652448987187803]
Knowing a news article's slant and authenticity is of crucial importance in times of "fake news"
We introduce Newsalyze, a bias-aware news reader focusing on a subtle, yet powerful form of media bias, named bias by word choice and labeling (WCL)
WCL bias can alter the assessment of entities reported in the news, e.g., "freedom fighters" vs. "terrorists"
arXiv Detail & Related papers (2021-05-20T11:20:37Z) - Misinfo Belief Frames: A Case Study on Covid & Climate News [49.979419711713795]
We propose a formalism for understanding how readers perceive the reliability of news and the impact of misinformation.
We introduce the Misinfo Belief Frames (MBF) corpus, a dataset of 66k inferences over 23.5k headlines.
Our results using large-scale language modeling to predict misinformation frames show that machine-generated inferences can influence readers' trust in news headlines.
arXiv Detail & Related papers (2021-04-18T09:50:11Z) - Viable Threat on News Reading: Generating Biased News Using Natural
Language Models [49.90665530780664]
We show that publicly available language models can reliably generate biased news content based on an input original news.
We also show that a large number of high-quality biased news articles can be generated using controllable text generation.
arXiv Detail & Related papers (2020-10-05T16:55:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.