On the Evaluation of the Plausibility and Faithfulness of Sentiment
Analysis Explanations
- URL: http://arxiv.org/abs/2210.06916v1
- Date: Thu, 13 Oct 2022 11:29:17 GMT
- Title: On the Evaluation of the Plausibility and Faithfulness of Sentiment
Analysis Explanations
- Authors: Julia El Zini, Mohamad Mansour, Basel Mousi, and Mariette Awad
- Abstract summary: We propose different metrics and techniques to evaluate the explainability of SA models from two angles.
First, we evaluate the strength of the extracted "rationales" in faithfully explaining the predicted outcome.
Second, we measure the agreement between ExAI methods and human judgment on a homegrown dataset.
- Score: 2.071923272918415
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Current Explainable AI (ExAI) methods, especially in the NLP field, are
conducted on various datasets by employing different metrics to evaluate
several aspects. The lack of a common evaluation framework is hindering the
progress tracking of such methods and their wider adoption. In this work,
inspired by offline information retrieval, we propose different metrics and
techniques to evaluate the explainability of SA models from two angles. First,
we evaluate the strength of the extracted "rationales" in faithfully explaining
the predicted outcome. Second, we measure the agreement between ExAI methods
and human judgment on a homegrown dataset1 to reflect on the rationales
plausibility. Our conducted experiments comprise four dimensions: (1) the
underlying architectures of SA models, (2) the approach followed by the ExAI
method, (3) the reasoning difficulty, and (4) the homogeneity of the
ground-truth rationales. We empirically demonstrate that anchors explanations
are more aligned with the human judgment and can be more confident in
extracting supporting rationales. As can be foreseen, the reasoning complexity
of sentiment is shown to thwart ExAI methods from extracting supporting
evidence. Moreover, a remarkable discrepancy is discerned between the results
of different explainability methods on the various architectures suggesting the
need for consolidation to observe enhanced performance. Predominantly,
transformers are shown to exhibit better explainability than convolutional and
recurrent architectures. Our work paves the way towards designing more
interpretable NLP models and enabling a common evaluation ground for their
relative strengths and robustness.
Related papers
- Learning to Generate and Evaluate Fact-checking Explanations with Transformers [10.970249299147866]
Research contributes to the field of Explainable Artificial Antelligence (XAI)
We develop transformer-based fact-checking models that contextualise and justify their decisions by generating human-accessible explanations.
We emphasise the need for aligning Artificial Intelligence (AI)-generated explanations with human judgements.
arXiv Detail & Related papers (2024-10-21T06:22:51Z) - Can you trust your explanations? A robustness test for feature attribution methods [42.36530107262305]
The field of Explainable AI (XAI) has seen a rapid growth but the usage of its techniques has at times led to unexpected results.
We will show how leveraging manifold hypothesis and ensemble approaches can be beneficial to an in-depth analysis of the robustness.
arXiv Detail & Related papers (2024-06-20T14:17:57Z) - A Survey of Explainable Knowledge Tracing [14.472784840283099]
This paper thoroughly analyzes the interpretability of KT algorithms.
Current evaluation methods for explainable knowledge tracing are lacking.
This paper offers some insights into evaluation methods from the perspective of educational stakeholders.
arXiv Detail & Related papers (2024-03-12T03:17:59Z) - Explaining Explainability: Towards Deeper Actionable Insights into Deep
Learning through Second-order Explainability [70.60433013657693]
Second-order explainable AI (SOXAI) was recently proposed to extend explainable AI (XAI) from the instance level to the dataset level.
We demonstrate for the first time, via example classification and segmentation cases, that eliminating irrelevant concepts from the training set based on actionable insights from SOXAI can enhance a model's performance.
arXiv Detail & Related papers (2023-06-14T23:24:01Z) - Multimodal Relation Extraction with Cross-Modal Retrieval and Synthesis [89.04041100520881]
This research proposes to retrieve textual and visual evidence based on the object, sentence, and whole image.
We develop a novel approach to synthesize the object-level, image-level, and sentence-level information for better reasoning between the same and different modalities.
arXiv Detail & Related papers (2023-05-25T15:26:13Z) - An Experimental Investigation into the Evaluation of Explainability
Methods [60.54170260771932]
This work compares 14 different metrics when applied to nine state-of-the-art XAI methods and three dummy methods (e.g., random saliency maps) used as references.
Experimental results show which of these metrics produces highly correlated results, indicating potential redundancy.
arXiv Detail & Related papers (2023-05-25T08:07:07Z) - INTERACTION: A Generative XAI Framework for Natural Language Inference
Explanations [58.062003028768636]
Current XAI approaches only focus on delivering a single explanation.
This paper proposes a generative XAI framework, INTERACTION (explaIn aNd predicT thEn queRy with contextuAl CondiTional varIational autO-eNcoder)
Our novel framework presents explanation in two steps: (step one) Explanation and Label Prediction; and (step two) Diverse Evidence Generation.
arXiv Detail & Related papers (2022-09-02T13:52:39Z) - Variational Distillation for Multi-View Learning [104.17551354374821]
We design several variational information bottlenecks to exploit two key characteristics for multi-view representation learning.
Under rigorously theoretical guarantee, our approach enables IB to grasp the intrinsic correlation between observations and semantic labels.
arXiv Detail & Related papers (2022-06-20T03:09:46Z) - Deep Context- and Relation-Aware Learning for Aspect-based Sentiment
Analysis [3.7175198778996483]
We propose Deep Contextualized Relation-Aware Network (DCRAN), which allows interactive relations among subtasks with deep contextual information.
DCRAN significantly outperforms previous state-of-the-art methods by large margins on three widely used benchmarks.
arXiv Detail & Related papers (2021-06-07T17:16:15Z) - A Diagnostic Study of Explainability Techniques for Text Classification [52.879658637466605]
We develop a list of diagnostic properties for evaluating existing explainability techniques.
We compare the saliency scores assigned by the explainability techniques with human annotations of salient input regions to find relations between a model's performance and the agreement of its rationales with human ones.
arXiv Detail & Related papers (2020-09-25T12:01:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.