M2P2: Multimodal Persuasion Prediction using Adaptive Fusion
- URL: http://arxiv.org/abs/2006.11405v2
- Date: Sat, 11 Dec 2021 20:50:15 GMT
- Title: M2P2: Multimodal Persuasion Prediction using Adaptive Fusion
- Authors: Chongyang Bai, Haipeng Chen, Srijan Kumar, Jure Leskovec, V.S.
Subrahmanian
- Abstract summary: This paper solves two problems: the Debate Outcome Prediction (DOP) problem predicts who wins a debate and the Intensity of Persuasion Prediction (IPP) problem predicts the change in the number of votes before and after a speaker speaks.
Our M2P2 framework is the first to use multimodal (acoustic, visual, language) data to solve the IPP problem.
- Score: 65.04045695380333
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Identifying persuasive speakers in an adversarial environment is a critical
task. In a national election, politicians would like to have persuasive
speakers campaign on their behalf. When a company faces adverse publicity, they
would like to engage persuasive advocates for their position in the presence of
adversaries who are critical of them. Debates represent a common platform for
these forms of adversarial persuasion. This paper solves two problems: the
Debate Outcome Prediction (DOP) problem predicts who wins a debate while the
Intensity of Persuasion Prediction (IPP) problem predicts the change in the
number of votes before and after a speaker speaks. Though DOP has been
previously studied, we are the first to study IPP. Past studies on DOP fail to
leverage two important aspects of multimodal data: 1) multiple modalities are
often semantically aligned, and 2) different modalities may provide diverse
information for prediction. Our M2P2 (Multimodal Persuasion Prediction)
framework is the first to use multimodal (acoustic, visual, language) data to
solve the IPP problem. To leverage the alignment of different modalities while
maintaining the diversity of the cues they provide, M2P2 devises a novel
adaptive fusion learning framework which fuses embeddings obtained from two
modules -- an alignment module that extracts shared information between
modalities and a heterogeneity module that learns the weights of different
modalities with guidance from three separately trained unimodal reference
models. We test M2P2 on the popular IQ2US dataset designed for DOP. We also
introduce a new dataset called QPS (from Qipashuo, a popular Chinese debate TV
show ) for IPP. M2P2 significantly outperforms 4 recent baselines on both
datasets.
Related papers
- Reasoning Beyond Bias: A Study on Counterfactual Prompting and Chain of Thought Reasoning [0.0]
We show that differences in learned regularities across answer options are predictive of model preferences and mirror human test-taking strategies.
We introduce two novel methods: Counterfactual Prompting with Chain of Thought (CoT) and Counterfactual Prompting with Agnostically Primed CoT (APriCoT)
Our results suggest that mitigating bias requires a "System-2" like process and that CoT reasoning is susceptible to confirmation bias under some prompting methodologies.
arXiv Detail & Related papers (2024-08-16T10:34:50Z) - ACQUIRED: A Dataset for Answering Counterfactual Questions In Real-Life
Videos [53.92440577914417]
ACQUIRED consists of 3.9K annotated videos, encompassing a wide range of event types and incorporating both first and third-person viewpoints.
Each video is annotated with questions that span three distinct dimensions of reasoning, including physical, social, and temporal.
We benchmark our dataset against several state-of-the-art language-only and multimodal models and experimental results demonstrate a significant performance gap.
arXiv Detail & Related papers (2023-11-02T22:17:03Z) - Detecting and Grounding Multi-Modal Media Manipulation and Beyond [93.08116982163804]
We highlight a new research problem for multi-modal fake media, namely Detecting and Grounding Multi-Modal Media Manipulation (DGM4)
DGM4 aims to not only detect the authenticity of multi-modal media, but also ground the manipulated content.
We propose a novel HierArchical Multi-modal Manipulation rEasoning tRansformer (HAMMER) to fully capture the fine-grained interaction between different modalities.
arXiv Detail & Related papers (2023-09-25T15:05:46Z) - GPTs at Factify 2022: Prompt Aided Fact-Verification [0.0]
We present our solution based on two approaches - PLM (pre-trained language model) based method and Prompt based method.
We achieved an F1 score of 0.6946 on the FACTIFY dataset and a 7th position on the competition leader-board.
arXiv Detail & Related papers (2022-06-29T21:07:39Z) - Visual Persuasion in COVID-19 Social Media Content: A Multi-Modal
Characterization [30.710295617831015]
This work proposes a computational approach to analyze the outcome of persuasive information in multi-modal content.
It focuses on two aspects, popularity and reliability, in COVID-19-related news articles shared on Twitter.
arXiv Detail & Related papers (2021-12-05T02:15:01Z) - Detecting Speaker Personas from Conversational Texts [52.4557098875992]
We study a new task, named Speaker Persona Detection (SPD), which aims to detect speaker personas based on the plain conversational text.
We build a dataset for SPD, dubbed as Persona Match on Persona-Chat (PMPC)
We evaluate several baseline models and propose utterance-to-profile (U2P) matching networks for this task.
arXiv Detail & Related papers (2021-09-03T06:14:38Z) - Knowledgeable Dialogue Reading Comprehension on Key Turns [84.1784903043884]
Multi-choice machine reading comprehension (MRC) requires models to choose the correct answer from candidate options given a passage and a question.
Our research focuses dialogue-based MRC, where the passages are multi-turn dialogues.
It suffers from two challenges, the answer selection decision is made without support of latently helpful commonsense, and the multi-turn context may hide considerable irrelevant information.
arXiv Detail & Related papers (2020-04-29T07:04:43Z) - DUMA: Reading Comprehension with Transposition Thinking [107.89721765056281]
Multi-choice Machine Reading (MRC) requires model to decide the correct answer from a set of answer options when given a passage and a question.
New DUal Multi-head Co-Attention (DUMA) model is inspired by human's transposition thinking process solving the multi-choice MRC problem.
arXiv Detail & Related papers (2020-01-26T07:35:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.