Interpretable Rumor Detection in Microblogs by Attending to User
Interactions
- URL: http://arxiv.org/abs/2001.10667v1
- Date: Wed, 29 Jan 2020 02:37:11 GMT
- Title: Interpretable Rumor Detection in Microblogs by Attending to User
Interactions
- Authors: Ling Min Serena Khoo, Hai Leong Chieu, Zhong Qian and Jing Jiang
- Abstract summary: We address rumor detection by learning to differentiate between the community's response to real and fake claims in microblogs.
We propose a post-level attention model (PLAN) to model long distance interactions between tweets.
We show that our best models outperform current state-of-the-art models for both data sets.
- Score: 11.881745037884953
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We address rumor detection by learning to differentiate between the
community's response to real and fake claims in microblogs. Existing
state-of-the-art models are based on tree models that model conversational
trees. However, in social media, a user posting a reply might be replying to
the entire thread rather than to a specific user. We propose a post-level
attention model (PLAN) to model long distance interactions between tweets with
the multi-head attention mechanism in a transformer network. We investigated
variants of this model: (1) a structure aware self-attention model (StA-PLAN)
that incorporates tree structure information in the transformer network, and
(2) a hierarchical token and post-level attention model (StA-HiTPLAN) that
learns a sentence representation with token-level self-attention. To the best
of our knowledge, we are the first to evaluate our models on two rumor
detection data sets: the PHEME data set as well as the Twitter15 and Twitter16
data sets. We show that our best models outperform current state-of-the-art
models for both data sets. Moreover, the attention mechanism allows us to
explain rumor detection predictions at both token-level and post-level.
Related papers
- Unidentified and Confounded? Understanding Two-Tower Models for Unbiased Learning to Rank [50.9530591265324]
Training two-tower models on clicks collected by well-performing production systems leads to decreased ranking performance.<n>We theoretically analyze the identifiability conditions of two-tower models, showing that either document swaps across positions or overlapping feature distributions are required to recover model parameters from clicks.<n>We also investigate the effect of logging policies on two-tower models, finding that they introduce no bias when models perfectly capture user behavior.
arXiv Detail & Related papers (2025-06-25T14:47:43Z) - FuXi-$α$: Scaling Recommendation Model with Feature Interaction Enhanced Transformer [81.12174905444229]
Recent advancements have shown that expanding sequential recommendation models to large-scale recommendation models can be an effective strategy.
We propose a new model called FuXi-$alpha$ to address these issues.
Our model outperforms existing models, with its performance continuously improving as the model size increases.
arXiv Detail & Related papers (2025-02-05T09:46:54Z) - Human-Object Interaction Detection Collaborated with Large Relation-driven Diffusion Models [65.82564074712836]
We introduce DIFfusionHOI, a new HOI detector shedding light on text-to-image diffusion models.
We first devise an inversion-based strategy to learn the expression of relation patterns between humans and objects in embedding space.
These learned relation embeddings then serve as textual prompts, to steer diffusion models generate images that depict specific interactions.
arXiv Detail & Related papers (2024-10-26T12:00:33Z) - NarrationDep: Narratives on Social Media For Automatic Depression Detection [24.11420537250414]
We have developed a novel model called textttNarrationDep, which focuses on detecting narratives associated with depression.
textttNarrationDep is a deep learning framework that jointly models individual user tweet representations and clusters of users' tweets.
arXiv Detail & Related papers (2024-07-24T11:24:25Z) - Evaluating Model Bias Requires Characterizing its Mistakes [19.777130236160712]
We introduce SkewSize, a principled and flexible metric that captures bias from mistakes in a model's predictions.
It can be used in multi-class settings or generalised to the open vocabulary setting of generative models.
We demonstrate the utility of SkewSize in multiple settings including: standard vision models trained on synthetic data, vision models trained on ImageNet, and large scale vision-and-language models from the BLIP-2 family.
arXiv Detail & Related papers (2024-07-15T11:46:21Z) - STMT: A Spatial-Temporal Mesh Transformer for MoCap-Based Action Recognition [50.064502884594376]
We study the problem of human action recognition using motion capture (MoCap) sequences.
We propose a novel Spatial-Temporal Mesh Transformer (STMT) to directly model the mesh sequences.
The proposed method achieves state-of-the-art performance compared to skeleton-based and point-cloud-based models.
arXiv Detail & Related papers (2023-03-31T16:19:27Z) - Large Language Models with Controllable Working Memory [64.71038763708161]
Large language models (LLMs) have led to a series of breakthroughs in natural language processing (NLP)
What further sets these models apart is the massive amounts of world knowledge they internalize during pretraining.
How the model's world knowledge interacts with the factual information presented in the context remains under explored.
arXiv Detail & Related papers (2022-11-09T18:58:29Z) - Analyzing Robustness of End-to-End Neural Models for Automatic Speech
Recognition [11.489161072526677]
We investigate robustness properties of pre-trained neural models for automatic speech recognition.
In this work, we perform a robustness analysis of the pre-trained neural models wav2vec2, HuBERT and DistilHuBERT on the LibriSpeech and TIMIT datasets.
arXiv Detail & Related papers (2022-08-17T20:00:54Z) - On the Role of Bidirectionality in Language Model Pre-Training [85.14614350372004]
We study the role of bidirectionality in next token prediction, text infilling, zero-shot priming and fine-tuning.
We train models with up to 6.7B parameters, and find differences to remain consistent at scale.
arXiv Detail & Related papers (2022-05-24T02:25:05Z) - Online Multiple Object Tracking with Cross-Task Synergy [120.70085565030628]
We propose a novel unified model with synergy between position prediction and embedding association.
The two tasks are linked by temporal-aware target attention and distractor attention, as well as identity-aware memory aggregation model.
arXiv Detail & Related papers (2021-04-01T10:19:40Z) - Probing Model Signal-Awareness via Prediction-Preserving Input
Minimization [67.62847721118142]
We evaluate models' ability to capture the correct vulnerability signals to produce their predictions.
We measure the signal awareness of models using a new metric we propose- Signal-aware Recall (SAR)
The results show a sharp drop in the model's Recall from the high 90s to sub-60s with the new metric.
arXiv Detail & Related papers (2020-11-25T20:05:23Z) - Rumor Detection on Twitter Using Multiloss Hierarchical BiLSTM with an
Attenuation Factor [14.717465036484292]
Social media platforms such as Twitter have become a breeding ground for unverified information or rumors.
Our model achieves better performance than that of state-of-the-art machine learning and vanilla deep learning models.
arXiv Detail & Related papers (2020-10-31T12:29:25Z) - On the Importance of Local Information in Transformer Based Models [19.036044858449593]
The self-attention module is a key component of Transformer-based models.
Recent studies have shown that these heads exhibit syntactic, semantic, or local behaviour.
We show that a larger fraction of heads have a locality bias as compared to a syntactic bias.
arXiv Detail & Related papers (2020-08-13T11:32:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.