Explaining Latent Representations with a Corpus of Examples
- URL: http://arxiv.org/abs/2110.15355v1
- Date: Thu, 28 Oct 2021 17:59:06 GMT
- Title: Explaining Latent Representations with a Corpus of Examples
- Authors: Jonathan Crabb\'e, Zhaozhi Qian, Fergus Imrie, Mihaela van der Schaar
- Abstract summary: We propose SimplEx: a user-centred method that provides example-based explanations with reference to a freely selected set of examples.
SimplEx uses the corpus to improve the user's understanding of the latent space with post-hoc explanations.
We show that SimplEx empowers the user by highlighting relevant patterns in the corpus that explain model representations.
- Score: 72.50996504722293
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern machine learning models are complicated. Most of them rely on
convoluted latent representations of their input to issue a prediction. To
achieve greater transparency than a black-box that connects inputs to
predictions, it is necessary to gain a deeper understanding of these latent
representations. To that aim, we propose SimplEx: a user-centred method that
provides example-based explanations with reference to a freely selected set of
examples, called the corpus. SimplEx uses the corpus to improve the user's
understanding of the latent space with post-hoc explanations answering two
questions: (1) Which corpus examples explain the prediction issued for a given
test example? (2) What features of these corpus examples are relevant for the
model to relate them to the test example? SimplEx provides an answer by
reconstructing the test latent representation as a mixture of corpus latent
representations. Further, we propose a novel approach, the Integrated Jacobian,
that allows SimplEx to make explicit the contribution of each corpus feature in
the mixture. Through experiments on tasks ranging from mortality prediction to
image classification, we demonstrate that these decompositions are robust and
accurate. With illustrative use cases in medicine, we show that SimplEx
empowers the user by highlighting relevant patterns in the corpus that explain
model representations. Moreover, we demonstrate how the freedom in choosing the
corpus allows the user to have personalized explanations in terms of examples
that are meaningful for them.
Related papers
- Answer is All You Need: Instruction-following Text Embedding via
Answering the Question [41.727700155498546]
This paper offers a new viewpoint, which treats the instruction as a question about the input text and encodes the expected answers to obtain the representation accordingly.
Specifically, we propose InBedder that instantiates this embed-via-answering idea by only fine-tuning language models on abstractive question answering tasks.
arXiv Detail & Related papers (2024-02-15T01:02:41Z) - Conjunct Resolution in the Face of Verbal Omissions [51.220650412095665]
We propose a conjunct resolution task that operates directly on the text and makes use of a split-and-rephrase paradigm in order to recover the missing elements in the coordination structure.
We curate a large dataset, containing over 10K examples of naturally-occurring verbal omissions with crowd-sourced annotations.
We train various neural baselines for this task, and show that while our best method obtains decent performance, it leaves ample space for improvement.
arXiv Detail & Related papers (2023-05-26T08:44:02Z) - Explanation Selection Using Unlabeled Data for Chain-of-Thought
Prompting [80.9896041501715]
Explanations that have not been "tuned" for a task, such as off-the-shelf explanations written by nonexperts, may lead to mediocre performance.
This paper tackles the problem of how to optimize explanation-infused prompts in a blackbox fashion.
arXiv Detail & Related papers (2023-02-09T18:02:34Z) - What Are You Token About? Dense Retrieval as Distributions Over the
Vocabulary [68.77983831618685]
We propose to interpret the vector representations produced by dual encoders by projecting them into the model's vocabulary space.
We show that the resulting projections contain rich semantic information, and draw connection between them and sparse retrieval.
arXiv Detail & Related papers (2022-12-20T16:03:25Z) - Contrastive Corpus Attribution for Explaining Representations [17.07084455770185]
Most explanation methods explain a scalar model output.
Recent works defined a scalar explanation output: a dot product-based similarity in the representation space to the sample being explained.
We propose contrastive corpus similarity, a novel and semantically meaningful scalar explanation output.
arXiv Detail & Related papers (2022-09-30T21:59:10Z) - Generating Informative Conclusions for Argumentative Texts [32.3103908466811]
The purpose of an argumentative text is to support a certain conclusion.
An explicit conclusion makes for a good candidate summary of an argumentative text.
This is especially true if the conclusion is informative, emphasizing specific concepts from the text.
arXiv Detail & Related papers (2021-06-02T10:35:59Z) - Contrastive Explanations for Model Interpretability [77.92370750072831]
We propose a methodology to produce contrastive explanations for classification models.
Our method is based on projecting model representation to a latent space.
Our findings shed light on the ability of label-contrastive explanations to provide a more accurate and finer-grained interpretability of a model's decision.
arXiv Detail & Related papers (2021-03-02T00:36:45Z) - Explanations of Black-Box Model Predictions by Contextual Importance and
Utility [1.7188280334580195]
We present the Contextual Importance (CI) and Contextual Utility (CU) concepts to extract explanations easily understandable by experts as well as novice users.
This method explains the prediction results without transforming the model into an interpretable one.
We show the utility of explanations in car selection example and Iris flower classification by presenting complete (i.e. the causes of an individual prediction) and contrastive explanation.
arXiv Detail & Related papers (2020-05-30T06:49:50Z) - Exploring Explainable Selection to Control Abstractive Summarization [51.74889133688111]
We develop a novel framework that focuses on explainability.
A novel pair-wise matrix captures the sentence interactions, centrality, and attribute scores.
A sentence-deployed attention mechanism in the abstractor ensures the final summary emphasizes the desired content.
arXiv Detail & Related papers (2020-04-24T14:39:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.