Enhancing Fake News Video Detection via LLM-Driven Creative Process Simulation
- URL: http://arxiv.org/abs/2510.04024v1
- Date: Sun, 05 Oct 2025 04:05:37 GMT
- Title: Enhancing Fake News Video Detection via LLM-Driven Creative Process Simulation
- Authors: Yuyan Bu, Qiang Sheng, Juan Cao, Shaofei Wang, Peng Qi, Yuhui Shi, Beizhe Hu,
- Abstract summary: The emergence of fake news on short video platforms has become a new significant societal concern.<n>Current detectors rely on pattern-based features to separate fake news videos from real ones.<n>We propose a data augmentation framework, AgentAug, that generates diverse fake news videos by simulating typical creative processes.
- Score: 14.79644134032037
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The emergence of fake news on short video platforms has become a new significant societal concern, necessitating automatic video-news-specific detection. Current detectors primarily rely on pattern-based features to separate fake news videos from real ones. However, limited and less diversified training data lead to biased patterns and hinder their performance. This weakness stems from the complex many-to-many relationships between video material segments and fabricated news events in real-world scenarios: a single video clip can be utilized in multiple ways to create different fake narratives, while a single fabricated event often combines multiple distinct video segments. However, existing datasets do not adequately reflect such relationships due to the difficulty of collecting and annotating large-scale real-world data, resulting in sparse coverage and non-comprehensive learning of the characteristics of potential fake news video creation. To address this issue, we propose a data augmentation framework, AgentAug, that generates diverse fake news videos by simulating typical creative processes. AgentAug implements multiple LLM-driven pipelines of four fabrication categories for news video creation, combined with an active learning strategy based on uncertainty sampling to select the potentially useful augmented samples during training. Experimental results on two benchmark datasets demonstrate that AgentAug consistently improves the performance of short video fake news detectors.
Related papers
- BindWeave: Subject-Consistent Video Generation via Cross-Modal Integration [56.98981194478512]
We propose a unified framework that handles a broad range of subject-to-video scenarios.<n>We introduce an MLLM-DiT framework in which a pretrained multimodal large language model performs deep cross-modal reasoning to ground entities.<n>Experiments on the OpenS2V benchmark demonstrate that our method achieves superior performance across subject consistency, naturalness, and text relevance in generated videos.
arXiv Detail & Related papers (2025-10-01T02:41:11Z) - Multimodal Learning for Fake News Detection in Short Videos Using Linguistically Verified Data and Heterogeneous Modality Fusion [5.850574227112314]
Current methods often struggle with the dynamic and multimodal nature of short video content.<n>This paper presents HFN, a novel framework that integrates video, audio, and text data to evaluate the authenticity of short video content.
arXiv Detail & Related papers (2025-09-19T04:24:57Z) - T$^\ ext{3}$SVFND: Towards an Evolving Fake News Detector for Emergencies with Test-time Training on Short Video Platforms [9.656794231914882]
We propose a new fake news videos detection framework (T$3$SVFND) using Test-Time Training (TTT)<n>Specifically, we design a self-supervised auxiliary task based on Mask Language Modeling (MLM) that masks a certain percentage of words in text.<n>In the test-time training phase, the model adapts to the distribution of test data through auxiliary tasks.
arXiv Detail & Related papers (2025-07-27T14:04:00Z) - FMNV: A Dataset of Media-Published News Videos for Fake News Detection [10.36393083923778]
We construct FMNV, a novel da-taset composed of news videos published by media organizations.<n>We employ Large Language Models (LLMs) to automatically generate content by manipulating authentic media-published news.<n>This work establishes critical benchmarks for de-tecting high-impact fake news in media ecosystems.
arXiv Detail & Related papers (2025-04-10T12:16:32Z) - WikiVideo: Article Generation from Multiple Videos [67.59430517160065]
We present the challenging task of automatically creating a high-level Wikipedia-style article that aggregates information from multiple videos about real-world events.<n>We introduce WikiVideo, a benchmark consisting of expert-written articles and densely annotated videos that provide evidence for articles' claims.<n>We propose Collaborative Article Generation (CAG), a novel interactive method for article creation from multiple videos.
arXiv Detail & Related papers (2025-04-01T16:22:15Z) - VMID: A Multimodal Fusion LLM Framework for Detecting and Identifying Misinformation of Short Videos [14.551693267228345]
This paper presents a novel fake news detection method based on multimodal information, designed to identify misinformation through a multi-level analysis of video content.
The proposed framework successfully integrates multimodal features within videos, significantly enhancing the accuracy and reliability of fake news detection.
arXiv Detail & Related papers (2024-11-15T08:20:26Z) - Official-NV: An LLM-Generated News Video Dataset for Multimodal Fake News Detection [9.48705939124715]
multimodal fake news detection has recently garnered increased attention.<n>We construct a dataset named Official-NV, comprising officially published news videos.<n>We also propose a new baseline model called OFNVD, which captures key information from multimodal features.
arXiv Detail & Related papers (2024-07-28T13:23:43Z) - FakingRecipe: Detecting Fake News on Short Video Platforms from the Perspective of Creative Process [19.629705422258905]
We introduce a novel perspective that considers how fake news might be created.
Through the lens of the creative process behind news video production, our empirical analysis uncovers the unique characteristics of fake news videos.
Based on the obtained insights, we design FakingRecipe, a creative process-aware model for detecting fake news short videos.
arXiv Detail & Related papers (2024-07-23T17:39:49Z) - VANE-Bench: Video Anomaly Evaluation Benchmark for Conversational LMMs [64.60035916955837]
VANE-Bench is a benchmark designed to assess the proficiency of Video-LMMs in detecting anomalies and inconsistencies in videos.<n>Our dataset comprises an array of videos synthetically generated using existing state-of-the-art text-to-video generation models.<n>We evaluate nine existing Video-LMMs, both open and closed sources, on this benchmarking task and find that most of the models encounter difficulties in effectively identifying the subtle anomalies.
arXiv Detail & Related papers (2024-06-14T17:59:01Z) - Multiverse: Multilingual Evidence for Fake News Detection [71.51905606492376]
Multiverse is a new feature based on multilingual evidence that can be used for fake news detection.
The hypothesis of the usage of cross-lingual evidence as a feature for fake news detection is confirmed.
arXiv Detail & Related papers (2022-11-25T18:24:17Z) - Towards Fast Adaptation of Pretrained Contrastive Models for
Multi-channel Video-Language Retrieval [70.30052749168013]
Multi-channel video-language retrieval require models to understand information from different channels.
contrastive multimodal models are shown to be highly effective at aligning entities in images/videos and text.
There is not a clear way to quickly adapt these two lines to multi-channel video-language retrieval with limited data and resources.
arXiv Detail & Related papers (2022-06-05T01:43:52Z) - Few-Shot Video Object Detection [70.43402912344327]
We introduce Few-Shot Video Object Detection (FSVOD) with three important contributions.
FSVOD-500 comprises of 500 classes with class-balanced videos in each category for few-shot learning.
Our TPN and TMN+ are jointly and end-to-end trained.
arXiv Detail & Related papers (2021-04-30T07:38:04Z) - Less is More: ClipBERT for Video-and-Language Learning via Sparse
Sampling [98.41300980759577]
A canonical approach to video-and-language learning dictates a neural model to learn from offline-extracted dense video features.
We propose a generic framework ClipBERT that enables affordable end-to-end learning for video-and-language tasks.
Experiments on text-to-video retrieval and video question answering on six datasets demonstrate that ClipBERT outperforms existing methods.
arXiv Detail & Related papers (2021-02-11T18:50:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.