Multi-modal Fake News Detection on Social Media via Multi-grained
Information Fusion
- URL: http://arxiv.org/abs/2304.00827v1
- Date: Mon, 3 Apr 2023 09:13:59 GMT
- Title: Multi-modal Fake News Detection on Social Media via Multi-grained
Information Fusion
- Authors: Yangming Zhou, Yuzhou Yang, Qichao Ying, Zhenxing Qian and Xinpeng
Zhang
- Abstract summary: We present a Multi-grained Multi-modal Fusion Network (MMFN) for fake news detection.
Inspired by the multi-grained process of human assessment of news authenticity, we respectively employ two Transformer-based pre-trained models to encode token-level features from text and images.
The multi-modal module fuses fine-grained features, taking into account coarse-grained features encoded by the CLIP encoder.
- Score: 21.042970740577648
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The easy sharing of multimedia content on social media has caused a rapid
dissemination of fake news, which threatens society's stability and security.
Therefore, fake news detection has garnered extensive research interest in the
field of social forensics. Current methods primarily concentrate on the
integration of textual and visual features but fail to effectively exploit
multi-modal information at both fine-grained and coarse-grained levels.
Furthermore, they suffer from an ambiguity problem due to a lack of correlation
between modalities or a contradiction between the decisions made by each
modality. To overcome these challenges, we present a Multi-grained Multi-modal
Fusion Network (MMFN) for fake news detection. Inspired by the multi-grained
process of human assessment of news authenticity, we respectively employ two
Transformer-based pre-trained models to encode token-level features from text
and images. The multi-modal module fuses fine-grained features, taking into
account coarse-grained features encoded by the CLIP encoder. To address the
ambiguity problem, we design uni-modal branches with similarity-based weighting
to adaptively adjust the use of multi-modal features. Experimental results
demonstrate that the proposed framework outperforms state-of-the-art methods on
three prevalent datasets.
Related papers
- Detecting Misinformation in Multimedia Content through Cross-Modal Entity Consistency: A Dual Learning Approach [10.376378437321437]
We propose a Multimedia Misinformation Detection framework for detecting misinformation from video content by leveraging cross-modal entity consistency.
Our results demonstrate that MultiMD outperforms state-of-the-art baseline models.
arXiv Detail & Related papers (2024-08-16T16:14:36Z) - Adaptive Learning of Consistency and Inconsistency Information for Fake News Detection [28.718460312783257]
We propose an adaptive multi-modal feature fusion network (MFF-Net) to detect fake news.
MFF-Net learns consistency information between modes through a multiple feature fusion module.
It outperforms state-of-the-art methods across three public news datasets derived from real social medias.
arXiv Detail & Related papers (2024-08-15T08:22:10Z) - Multi-modal Crowd Counting via a Broker Modality [64.5356816448361]
Multi-modal crowd counting involves estimating crowd density from both visual and thermal/depth images.
We propose a novel approach by introducing an auxiliary broker modality and frame the task as a triple-modal learning problem.
We devise a fusion-based method to generate this broker modality, leveraging a non-diffusion, lightweight counterpart of modern denoising diffusion-based fusion models.
arXiv Detail & Related papers (2024-07-10T10:13:11Z) - Multimodal Fusion on Low-quality Data: A Comprehensive Survey [110.22752954128738]
This paper surveys the common challenges and recent advances of multimodal fusion in the wild.
We identify four main challenges that are faced by multimodal fusion on low-quality data.
This new taxonomy will enable researchers to understand the state of the field and identify several potential directions.
arXiv Detail & Related papers (2024-04-27T07:22:28Z) - Detecting and Grounding Multi-Modal Media Manipulation and Beyond [93.08116982163804]
We highlight a new research problem for multi-modal fake media, namely Detecting and Grounding Multi-Modal Media Manipulation (DGM4)
DGM4 aims to not only detect the authenticity of multi-modal media, but also ground the manipulated content.
We propose a novel HierArchical Multi-modal Manipulation rEasoning tRansformer (HAMMER) to fully capture the fine-grained interaction between different modalities.
arXiv Detail & Related papers (2023-09-25T15:05:46Z) - Inconsistent Matters: A Knowledge-guided Dual-consistency Network for
Multi-modal Rumor Detection [53.48346699224921]
A novel Knowledge-guided Dualconsistency Network is proposed to detect rumors with multimedia contents.
It uses two consistency detectionworks to capture the inconsistency at the cross-modal level and the content-knowledge level simultaneously.
It also enables robust multi-modal representation learning under different missing visual modality conditions.
arXiv Detail & Related papers (2023-06-03T15:32:20Z) - Provable Dynamic Fusion for Low-Quality Multimodal Data [94.39538027450948]
Dynamic multimodal fusion emerges as a promising learning paradigm.
Despite its widespread use, theoretical justifications in this field are still notably lacking.
This paper provides theoretical understandings to answer this question under a most popular multimodal fusion framework from the generalization perspective.
A novel multimodal fusion framework termed Quality-aware Multimodal Fusion (QMF) is proposed, which can improve the performance in terms of classification accuracy and model robustness.
arXiv Detail & Related papers (2023-06-03T08:32:35Z) - Cross-modal Contrastive Learning for Multimodal Fake News Detection [10.760000041969139]
COOLANT is a cross-modal contrastive learning framework for multimodal fake news detection.
A cross-modal fusion module is developed to learn the cross-modality correlations.
An attention guidance module is implemented to help effectively and interpretably aggregate the aligned unimodal representations.
arXiv Detail & Related papers (2023-02-25T10:12:34Z) - Multimodal Fake News Detection with Adaptive Unimodal Representation
Aggregation [28.564442206829625]
AURA is a multimodal fake news detection network with adaptive unimodal representation aggregation.
We perform coarse-level fake news detection and cross-modal cosistency learning according to the unimodal and multimodal representations.
Experiments on Weibo and Gossipcop prove that AURA can successfully beat several state-of-the-art FND schemes.
arXiv Detail & Related papers (2022-06-12T14:06:55Z) - Multimodal Categorization of Crisis Events in Social Media [81.07061295887172]
We present a new multimodal fusion method that leverages both images and texts as input.
In particular, we introduce a cross-attention module that can filter uninformative and misleading components from weak modalities.
We show that our method outperforms the unimodal approaches and strong multimodal baselines by a large margin on three crisis-related tasks.
arXiv Detail & Related papers (2020-04-10T06:31:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.