Robust Misinformation Detection by Visiting Potential Commonsense Conflict
- URL: http://arxiv.org/abs/2504.21604v1
- Date: Wed, 30 Apr 2025 13:03:17 GMT
- Title: Robust Misinformation Detection by Visiting Potential Commonsense Conflict
- Authors: Bing Wang, Ximing Li, Changchun Li, Bingrui Zhao, Bo Fu, Renchu Guan, Shengsheng Wang,
- Abstract summary: Misinformation Detection (MD) aims to detect online misinformation automatically.<n>We propose a novel plug-and-play augmentation method for the MD task, namely Misinformation Detection with Potential Commonsense Conflict.<n>We take inspiration from the prior studies indicating that fake articles are more likely to involve commonsense conflict.
- Score: 25.27903714824545
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The development of Internet technology has led to an increased prevalence of misinformation, causing severe negative effects across diverse domains. To mitigate this challenge, Misinformation Detection (MD), aiming to detect online misinformation automatically, emerges as a rapidly growing research topic in the community. In this paper, we propose a novel plug-and-play augmentation method for the MD task, namely Misinformation Detection with Potential Commonsense Conflict (MD-PCC). We take inspiration from the prior studies indicating that fake articles are more likely to involve commonsense conflict. Accordingly, we construct commonsense expressions for articles, serving to express potential commonsense conflicts inferred by the difference between extracted commonsense triplet and golden ones inferred by the well-established commonsense reasoning tool COMET. These expressions are then specified for each article as augmentation. Any specific MD methods can be then trained on those commonsense-augmented articles. Besides, we also collect a novel commonsense-oriented dataset named CoMis, whose all fake articles are caused by commonsense conflict. We integrate MD-PCC with various existing MD backbones and compare them across both 4 public benchmark datasets and CoMis. Empirical results demonstrate that MD-PCC can consistently outperform the existing MD baselines.
Related papers
- Retrieval-Augmented Generation with Conflicting Evidence [57.66282463340297]
Large language model (LLM) agents are increasingly employing retrieval-augmented generation (RAG) to improve the factuality of their responses.<n>In practice, these systems often need to handle ambiguous user queries and potentially conflicting information from multiple sources.<n>We propose RAMDocs (Retrieval with Ambiguity and Misinformation in Documents), a new dataset that simulates complex and realistic scenarios for conflicting evidence for a user query.
arXiv Detail & Related papers (2025-04-17T16:46:11Z) - Pub-Guard-LLM: Detecting Fraudulent Biomedical Articles with Reliable Explanations [11.082285990214595]
Pub-Guard-LLM is a large language model-based system tailored to fraud detection of biomedical scientific articles.<n>Pub-Guard-LLM consistently surpasses the performance of various baselines.<n>By enhancing both detection performance and explainability in scientific fraud detection, Pub-Guard-LLM contributes to safeguarding research integrity with a novel, effective, open-source tool.
arXiv Detail & Related papers (2025-02-21T12:54:56Z) - What Really is Commonsense Knowledge? [58.5342212738895]
We survey existing definitions of commonsense knowledge, ground into the three frameworks for defining concepts, and consolidate them into a unified definition of commonsense knowledge.
We then use the consolidated definition for annotations and experiments on the CommonsenseQA and CommonsenseQA 2.0 datasets.
Our study shows that there exists a large portion of non-commonsense-knowledge instances in the two datasets, and a large performance gap on these two subsets.
arXiv Detail & Related papers (2024-11-06T14:54:19Z) - Why Misinformation is Created? Detecting them by Integrating Intent Features [25.20744191980224]
Social media platforms allow people to disseminate a plethora of information more efficiently and conveniently.
They are inevitably full of misinformation, causing damage to diverse aspects of our daily lives.
Misinformation Detection (MD) has become an active research topic receiving widespread attention.
arXiv Detail & Related papers (2024-07-27T07:30:47Z) - Harmfully Manipulated Images Matter in Multimodal Misinformation Detection [22.236455110413264]
Multimodal Misinformation Detection (MMD) has attracted growing attention from the academic and industrial communities.
We propose a novel HAMI-M3D method, namely Harmfully Manipulated Images Matter in MMD (HAMI-M3D)
Extensive experiments across three benchmark datasets can demonstrate that HAMI-M3D can consistently improve the performance of any MMD baselines.
arXiv Detail & Related papers (2024-07-27T07:16:07Z) - MMFakeBench: A Mixed-Source Multimodal Misinformation Detection Benchmark for LVLMs [47.353720361676004]
multimodal misinformation detection methods often assume a single source and type of forgery for each sample.<n>The lack of a benchmark for mixed-source misinformation has hindered progress in this field.<n> MMFakeBench is the first comprehensive benchmark for mixed-source MMD.
arXiv Detail & Related papers (2024-06-13T03:04:28Z) - Detecting and Grounding Multi-Modal Media Manipulation and Beyond [93.08116982163804]
We highlight a new research problem for multi-modal fake media, namely Detecting and Grounding Multi-Modal Media Manipulation (DGM4)
DGM4 aims to not only detect the authenticity of multi-modal media, but also ground the manipulated content.
We propose a novel HierArchical Multi-modal Manipulation rEasoning tRansformer (HAMMER) to fully capture the fine-grained interaction between different modalities.
arXiv Detail & Related papers (2023-09-25T15:05:46Z) - DCID: Deep Canonical Information Decomposition [84.59396326810085]
We consider the problem of identifying the signal shared between two one-dimensional target variables.
We propose ICM, an evaluation metric which can be used in the presence of ground-truth labels.
We also propose Deep Canonical Information Decomposition (DCID) - a simple, yet effective approach for learning the shared variables.
arXiv Detail & Related papers (2023-06-27T16:59:06Z) - ManiTweet: A New Benchmark for Identifying Manipulation of News on Social Media [74.93847489218008]
We present a novel task, identifying manipulation of news on social media, which aims to detect manipulation in social media posts and identify manipulated or inserted information.<n>To study this task, we have proposed a data collection schema and curated a dataset called ManiTweet, consisting of 3.6K pairs of tweets and corresponding articles.<n>Our analysis demonstrates that this task is highly challenging, with large language models (LLMs) yielding unsatisfactory performance.
arXiv Detail & Related papers (2023-05-23T16:40:07Z) - Synthetic Misinformers: Generating and Combating Multimodal
Misinformation [11.696058634552147]
multimodal misinformation detection (MMD) detects whether the combination of an image and its accompanying text could mislead or misinform.
We show that our proposed CLIP-based Named Entity Swapping can lead to MMD models that surpass other OOC and NEI Misinformers in terms of multimodal accuracy.
arXiv Detail & Related papers (2023-03-02T12:59:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.