Analyzing Modality Robustness in Multimodal Sentiment Analysis
- URL: http://arxiv.org/abs/2205.15465v1
- Date: Mon, 30 May 2022 23:30:16 GMT
- Title: Analyzing Modality Robustness in Multimodal Sentiment Analysis
- Authors: Devamanyu Hazarika, Yingting Li, Bo Cheng, Shuai Zhao, Roger
Zimmermann, Soujanya Poria
- Abstract summary: Building robust multimodal models is crucial for achieving reliable deployment in the wild.
We propose simple diagnostic checks for modality robustness in a trained multimodal model.
We analyze well-known robust training strategies to alleviate the issues.
- Score: 48.52878002917685
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Building robust multimodal models are crucial for achieving reliable
deployment in the wild. Despite its importance, less attention has been paid to
identifying and improving the robustness of Multimodal Sentiment Analysis (MSA)
models. In this work, we hope to address that by (i) Proposing simple
diagnostic checks for modality robustness in a trained multimodal model. Using
these checks, we find MSA models to be highly sensitive to a single modality,
which creates issues in their robustness; (ii) We analyze well-known robust
training strategies to alleviate the issues. Critically, we observe that
robustness can be achieved without compromising on the original performance. We
hope our extensive study-performed across five models and two benchmark
datasets-and proposed procedures would make robustness an integral component in
MSA research. Our diagnostic checks and robust training solutions are simple to
implement and available at https://github. com/declare-lab/MSA-Robustness.
Related papers
- RADAR: Robust Two-stage Modality-incomplete Industrial Anomaly Detection [61.71770293720491]
We propose a novel two-stage Robust modAlity-imcomplete fusing and Detecting frAmewoRk, abbreviated as RADAR.
Our bootstrapping philosophy is to enhance two stages in MIIAD, improving the robustness of the Multimodal Transformer.
Our experimental results demonstrate that the proposed RADAR significantly surpasses conventional MIAD methods in terms of effectiveness and robustness.
arXiv Detail & Related papers (2024-10-02T16:47:55Z) - Multimodal Large Language Models to Support Real-World Fact-Checking [80.41047725487645]
Multimodal large language models (MLLMs) carry the potential to support humans in processing vast amounts of information.
While MLLMs are already being used as a fact-checking tool, their abilities and limitations in this regard are understudied.
We propose a framework for systematically assessing the capacity of current multimodal models to facilitate real-world fact-checking.
arXiv Detail & Related papers (2024-03-06T11:32:41Z) - Quantifying and Enhancing Multi-modal Robustness with Modality Preference [9.367733452960492]
Multi-modal models are vulnerable to pervasive perturbations, such as uni-modal attacks and missing conditions.
Larger uni-modal representation margins and more reliable integration for modalities are essential components for achieving higher robustness.
Inspired by our theoretical finding, we introduce a training procedure called Certifiable Robust Multi-modal Training.
arXiv Detail & Related papers (2024-02-09T08:33:48Z) - RoAST: Robustifying Language Models via Adversarial Perturbation with
Selective Training [105.02614392553198]
We propose Robustifying LMs via Adversarial perturbation with Selective Training (RoAST)
RoAST incorporates two important sources for the model robustness, robustness on the perturbed inputs and generalizable knowledge in pre-trained LMs.
We demonstrate the effectiveness of RoAST compared to state-of-the-art fine-tuning methods on six different types of LMs.
arXiv Detail & Related papers (2023-12-07T04:23:36Z) - A Comprehensive Study on Robustness of Image Classification Models:
Benchmarking and Rethinking [54.89987482509155]
robustness of deep neural networks is usually lacking under adversarial examples, common corruptions, and distribution shifts.
We establish a comprehensive benchmark robustness called textbfARES-Bench on the image classification task.
By designing the training settings accordingly, we achieve the new state-of-the-art adversarial robustness.
arXiv Detail & Related papers (2023-02-28T04:26:20Z) - Defending Multimodal Fusion Models against Single-Source Adversaries [6.019777076722421]
We show that standard multimodal fusion models are vulnerable to single-source adversaries.
An attack on any single modality can overcome the correct information from multiple unperturbed modalities and cause the model to fail.
Motivated by this finding, we propose an adversarially robust fusion strategy.
arXiv Detail & Related papers (2022-06-25T18:57:02Z) - Measure and Improve Robustness in NLP Models: A Survey [23.515869499536237]
robustness has been separately explored in applications like vision and NLP, with various definitions, evaluation and mitigation strategies in multiple lines of research.
We first connect multiple definitions of robustness, then unify various lines of work on identifying robustness failures and evaluating models' robustness.
We present mitigation strategies that are data-driven, model-driven, and inductive-prior-based, with a more systematic view of how to effectively improve robustness in NLP models.
arXiv Detail & Related papers (2021-12-15T18:02:04Z) - On Fast Adversarial Robustness Adaptation in Model-Agnostic
Meta-Learning [100.14809391594109]
Model-agnostic meta-learning (MAML) has emerged as one of the most successful meta-learning techniques in few-shot learning.
Despite the generalization power of the meta-model, it remains elusive that how adversarial robustness can be maintained by MAML in few-shot learning.
We propose a general but easily-optimized robustness-regularized meta-learning framework, which allows the use of unlabeled data augmentation, fast adversarial attack generation, and computationally-light fine-tuning.
arXiv Detail & Related papers (2021-02-20T22:03:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.