Words are the Window to the Soul: Language-based User Representations
for Fake News Detection
- URL: http://arxiv.org/abs/2011.07389v1
- Date: Sat, 14 Nov 2020 21:14:17 GMT
- Title: Words are the Window to the Soul: Language-based User Representations
for Fake News Detection
- Authors: Marco Del Tredici and Raquel Fern\'andez
- Abstract summary: We introduce a model that creates representations of individuals on social media based only on the language they produce.
We show that language-based user representations are beneficial for this task.
- Score: 5.876243339384605
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Cognitive and social traits of individuals are reflected in language use.
Moreover, individuals who are prone to spread fake news online often share
common traits. Building on these ideas, we introduce a model that creates
representations of individuals on social media based only on the language they
produce, and use them to detect fake news. We show that language-based user
representations are beneficial for this task. We also present an extended
analysis of the language of fake news spreaders, showing that its main features
are mostly domain independent and consistent across two English datasets.
Finally, we exploit the relation between language use and connections in the
social graph to assess the presence of the Echo Chamber effect in our data.
Related papers
- Which linguistic cues make people fall for fake news? A comparison of
cognitive and affective processing [21.881235152669564]
Linguistic cues (e.g. adverbs, personal pronouns, positive emotion words, negative emotion words) are important characteristics of any text.
We compare the role of linguistic cues across both cognitive processing (related to careful thinking) and affective processing (related to unconscious automatic evaluations)
We find that users engage more in cognitive processing for longer fake news articles, while affective processing is more pronounced for fake news written in analytic words.
arXiv Detail & Related papers (2023-12-02T11:06:14Z) - To Tell The Truth: Language of Deception and Language Models [6.80186731352488]
We analyze a novel TV game show data where conversations in a high-stake environment result in lies.
We investigate the manifestation of potentially verifiable language cues of deception in the presence of objective truth.
We show that there exists a class of detectors (algorithms) that have similar truth detection performance compared to human subjects.
arXiv Detail & Related papers (2023-11-13T05:40:11Z) - Decoding the Silent Majority: Inducing Belief Augmented Social Graph
with Large Language Model for Response Forecasting [74.68371461260946]
SocialSense is a framework that induces a belief-centered graph on top of an existent social network, along with graph-based propagation to capture social dynamics.
Our method surpasses existing state-of-the-art in experimental evaluations for both zero-shot and supervised settings.
arXiv Detail & Related papers (2023-10-20T06:17:02Z) - Learning Cross-lingual Visual Speech Representations [108.68531445641769]
Cross-lingual self-supervised visual representation learning has been a growing research topic in the last few years.
We use the recently-proposed Raw Audio-Visual Speechs (RAVEn) framework to pre-train an audio-visual model with unlabelled data.
Our experiments show that: (1) multi-lingual models with more data outperform monolingual ones, but, when keeping the amount of data fixed, monolingual models tend to reach better performance.
arXiv Detail & Related papers (2023-03-14T17:05:08Z) - Multiverse: Multilingual Evidence for Fake News Detection [71.51905606492376]
Multiverse is a new feature based on multilingual evidence that can be used for fake news detection.
The hypothesis of the usage of cross-lingual evidence as a feature for fake news detection is confirmed.
arXiv Detail & Related papers (2022-11-25T18:24:17Z) - Detecting Offensive Language on Social Networks: An End-to-end Detection
Method based on Graph Attention Networks [7.723697303436006]
We propose an end-to-end method based on community structure and text features for offensive language detection (CT-OLD)
We add user opinion to the community structure for representing user features. The user opinion is represented by user historical behavior information, which outperforms that represented by text information.
arXiv Detail & Related papers (2022-03-04T03:57:18Z) - A New Generation of Perspective API: Efficient Multilingual
Character-level Transformers [66.9176610388952]
We present the fundamentals behind the next version of the Perspective API from Google Jigsaw.
At the heart of the approach is a single multilingual token-free Charformer model.
We demonstrate that by forgoing static vocabularies, we gain flexibility across a variety of settings.
arXiv Detail & Related papers (2022-02-22T20:55:31Z) - Bridging the Gap: Using Deep Acoustic Representations to Learn Grounded
Language from Percepts and Raw Speech [26.076534338576234]
Learning to understand grounded language, which connects natural language to percepts, is a critical research area.
In this work we demonstrate the feasibility of performing grounded language acquisition on paired visual percepts and raw speech inputs.
arXiv Detail & Related papers (2021-12-27T16:12:30Z) - Can You be More Social? Injecting Politeness and Positivity into
Task-Oriented Conversational Agents [60.27066549589362]
Social language used by human agents is associated with greater users' responsiveness and task completion.
The model uses a sequence-to-sequence deep learning architecture, extended with a social language understanding element.
Evaluation in terms of content preservation and social language level using both human judgment and automatic linguistic measures shows that the model can generate responses that enable agents to address users' issues in a more socially appropriate way.
arXiv Detail & Related papers (2020-12-29T08:22:48Z) - Vokenization: Improving Language Understanding with Contextualized,
Visual-Grounded Supervision [110.66085917826648]
We develop a technique that extrapolates multimodal alignments to language-only data by contextually mapping language tokens to their related images.
"vokenization" is trained on relatively small image captioning datasets and we then apply it to generate vokens for large language corpora.
Trained with these contextually generated vokens, our visually-supervised language models show consistent improvements over self-supervised alternatives on multiple pure-language tasks.
arXiv Detail & Related papers (2020-10-14T02:11:51Z) - Compositional Demographic Word Embeddings [41.89745054269992]
We propose a new form of personalized word embeddings that use demographic-specific word representations derived compositionally from full or partial demographic information for a user.
We show that the resulting demographic-aware word representations outperform generic word representations on two tasks for English: language modeling and word associations.
arXiv Detail & Related papers (2020-10-06T19:23:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.