Measuring Information in Text Explanations
- URL: http://arxiv.org/abs/2310.04557v1
- Date: Fri, 6 Oct 2023 19:46:51 GMT
- Title: Measuring Information in Text Explanations
- Authors: Zining Zhu, Frank Rudzicz
- Abstract summary: We argue that placing the explanations on an information-theoretic framework could unify the evaluations of two popular text explanation methods.
We quantify the information flow through these channels, thereby facilitating the assessment of explanation characteristics.
Our work contributes to the ongoing efforts in establishing rigorous and standardized evaluation criteria in the rapidly evolving field of explainable AI.
- Score: 23.929076318334047
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Text-based explanation is a particularly promising approach in explainable
AI, but the evaluation of text explanations is method-dependent. We argue that
placing the explanations on an information-theoretic framework could unify the
evaluations of two popular text explanation methods: rationale and natural
language explanations (NLE). This framework considers the post-hoc text
pipeline as a series of communication channels, which we refer to as
``explanation channels''. We quantify the information flow through these
channels, thereby facilitating the assessment of explanation characteristics.
We set up tools for quantifying two information scores: relevance and
informativeness. We illustrate what our proposed information scores measure by
comparing them against some traditional evaluation metrics. Our
information-theoretic scores reveal some unique observations about the
underlying mechanisms of two representative text explanations. For example, the
NLEs trade-off slightly between transmitting the input-related information and
the target-related information, whereas the rationales do not exhibit such a
trade-off mechanism. Our work contributes to the ongoing efforts in
establishing rigorous and standardized evaluation criteria in the rapidly
evolving field of explainable AI.
Related papers
- Explaining Explainability: Towards Deeper Actionable Insights into Deep
Learning through Second-order Explainability [70.60433013657693]
Second-order explainable AI (SOXAI) was recently proposed to extend explainable AI (XAI) from the instance level to the dataset level.
We demonstrate for the first time, via example classification and segmentation cases, that eliminating irrelevant concepts from the training set based on actionable insights from SOXAI can enhance a model's performance.
arXiv Detail & Related papers (2023-06-14T23:24:01Z) - Counterfactuals of Counterfactuals: a back-translation-inspired approach
to analyse counterfactual editors [3.4253416336476246]
We focus on the analysis of counterfactual, contrastive explanations.
We propose a new back translation-inspired evaluation methodology.
We show that by iteratively feeding the counterfactual to the explainer we can obtain valuable insights into the behaviour of both the predictor and the explainer models.
arXiv Detail & Related papers (2023-05-26T16:04:28Z) - Evaluation of Automatically Constructed Word Meaning Explanations [0.0]
We present a new tool that derives explanations automatically based on collective information from very large corpora.
We show that the presented approach allows to create explanations that contain data useful for understanding the word meaning in approximately 90% of cases.
arXiv Detail & Related papers (2023-02-27T09:47:55Z) - Explanation Selection Using Unlabeled Data for Chain-of-Thought
Prompting [80.9896041501715]
Explanations that have not been "tuned" for a task, such as off-the-shelf explanations written by nonexperts, may lead to mediocre performance.
This paper tackles the problem of how to optimize explanation-infused prompts in a blackbox fashion.
arXiv Detail & Related papers (2023-02-09T18:02:34Z) - REVEL Framework to measure Local Linear Explanations for black-box
models: Deep Learning Image Classification case of study [12.49538398746092]
We propose a procedure called REVEL to evaluate different aspects concerning the quality of explanations with a theoretically coherent development.
The experiments have been carried out on image four datasets as benchmark where we show REVEL's descriptive and analytical power.
arXiv Detail & Related papers (2022-11-11T12:15:36Z) - REV: Information-Theoretic Evaluation of Free-Text Rationales [83.24985872655738]
We argue that an ideal metric should focus on the new information uniquely provided in the rationale that is otherwise not provided in the input or the label.
We propose a metric called REV (Rationale Evaluation with conditional V-information) to quantify the amount of new, label-relevant information in a rationale.
arXiv Detail & Related papers (2022-10-10T19:31:30Z) - Logical Satisfiability of Counterfactuals for Faithful Explanations in
NLI [60.142926537264714]
We introduce the methodology of Faithfulness-through-Counterfactuals.
It generates a counterfactual hypothesis based on the logical predicates expressed in the explanation.
It then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic.
arXiv Detail & Related papers (2022-05-25T03:40:59Z) - The Unreliability of Explanations in Few-Shot In-Context Learning [50.77996380021221]
We focus on two NLP tasks that involve reasoning over text, namely question answering and natural language inference.
We show that explanations judged as good by humans--those that are logically consistent with the input--usually indicate more accurate predictions.
We present a framework for calibrating model predictions based on the reliability of the explanations.
arXiv Detail & Related papers (2022-05-06T17:57:58Z) - Human Interpretation of Saliency-based Explanation Over Text [65.29015910991261]
We study saliency-based explanations over textual data.
We find that people often mis-interpret the explanations.
We propose a method to adjust saliencies based on model estimates of over- and under-perception.
arXiv Detail & Related papers (2022-01-27T15:20:32Z) - Uncertainty Quantification of Surrogate Explanations: an Ordinal
Consensus Approach [1.3750624267664155]
We produce estimates of the uncertainty of a given explanation by measuring the consensus amongst a set of diverse bootstrapped surrogate explainers.
We empirically illustrate the properties of this approach through experiments on state-of-the-art Convolutional Neural Network ensembles.
arXiv Detail & Related papers (2021-11-17T13:55:58Z) - The Explanation Game: Towards Prediction Explainability through Sparse
Communication [6.497816402045099]
We provide a unified perspective of explainability as a problem between an explainer and a layperson.
We use this framework to compare several prior approaches for extracting explanations.
We propose new embedded methods for explainability, through the use of selective, sparse attention.
arXiv Detail & Related papers (2020-04-28T22:27:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.