LEx: A Framework for Operationalising Layers of Machine Learning
Explanations
- URL: http://arxiv.org/abs/2104.09612v1
- Date: Thu, 15 Apr 2021 23:31:04 GMT
- Title: LEx: A Framework for Operationalising Layers of Machine Learning
Explanations
- Authors: Ronal Singh, Upol Ehsan, Marc Cheong, Mark O. Riedl, Tim Miller
- Abstract summary: We define a framework called the textitlayers of explanation (LEx)
The framework uses the notions of textitsensitivity (emotional responsiveness) of features and the level of textitstakes (decision's consequence) in a domain to determine whether different types of explanations are textitappropriate in a given context.
We demonstrate how to use the framework to assess the appropriateness of different types of explanations in different domains.
- Score: 10.46427915540552
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Several social factors impact how people respond to AI explanations used to
justify AI decisions affecting them personally. In this position paper, we
define a framework called the \textit{layers of explanation} (LEx), a lens
through which we can assess the appropriateness of different types of
explanations. The framework uses the notions of \textit{sensitivity} (emotional
responsiveness) of features and the level of \textit{stakes} (decision's
consequence) in a domain to determine whether different types of explanations
are \textit{appropriate} in a given context. We demonstrate how to use the
framework to assess the appropriateness of different types of explanations in
different domains.
Related papers
- An Ontology-Enabled Approach For User-Centered and Knowledge-Enabled Explanations of AI Systems [0.3480973072524161]
Recent research in explainability has focused on explaining the workings of AI models or model explainability.
This thesis seeks to bridge some gaps between model and user-centered explainability.
arXiv Detail & Related papers (2024-10-23T02:03:49Z) - Fool Me Once? Contrasting Textual and Visual Explanations in a Clinical Decision-Support Setting [43.110187812734864]
We evaluate three types of explanations: visual explanations (saliency maps), natural language explanations, and a combination of both modalities.
We find that text-based explanations lead to significant over-reliance, which is alleviated by combining them with saliency maps.
We also observe that the quality of explanations, that is, how much factually correct information they entail, and how much this aligns with AI correctness, significantly impacts the usefulness of the different explanation types.
arXiv Detail & Related papers (2024-10-16T06:43:02Z) - The Shape of Explanations: A Topological Account of Rule-Based
Explanations in Machine Learning [0.0]
We introduce a framework for rule-based explanation methods and provide a characterization of explainability.
We argue that the preferred scheme depends on how much the user knows about the domain and the probability measure over the feature space.
arXiv Detail & Related papers (2023-01-22T02:58:00Z) - Complementary Explanations for Effective In-Context Learning [77.83124315634386]
Large language models (LLMs) have exhibited remarkable capabilities in learning from explanations in prompts.
This work aims to better understand the mechanisms by which explanations are used for in-context learning.
arXiv Detail & Related papers (2022-11-25T04:40:47Z) - Sensible AI: Re-imagining Interpretability and Explainability using
Sensemaking Theory [14.35488479818285]
We propose an alternate framework for interpretability grounded in Weick's sensemaking theory.
We use an application of sensemaking in organizations as a template for discussing design guidelines for Sensible AI.
arXiv Detail & Related papers (2022-05-10T17:20:44Z) - Do Explanations Explain? Model Knows Best [39.86131552976105]
It is a mystery which input features contribute to a neural network's output.
We propose a framework for evaluating the explanations using the neural network model itself.
arXiv Detail & Related papers (2022-03-04T12:39:29Z) - Human Interpretation of Saliency-based Explanation Over Text [65.29015910991261]
We study saliency-based explanations over textual data.
We find that people often mis-interpret the explanations.
We propose a method to adjust saliencies based on model estimates of over- and under-perception.
arXiv Detail & Related papers (2022-01-27T15:20:32Z) - Explanatory Pluralism in Explainable AI [0.0]
I chart a taxonomy of types of explanation and the associated XAI methods that can address them.
When we look to expose the inner mechanisms of AI models, we produce Diagnostic-explanations.
When we wish to form stable generalizations of our models, we produce Expectation-explanations.
Finally, when we want to justify the usage of a model, we produce Role-explanations.
arXiv Detail & Related papers (2021-06-26T09:02:06Z) - Contrastive Explanations for Model Interpretability [77.92370750072831]
We propose a methodology to produce contrastive explanations for classification models.
Our method is based on projecting model representation to a latent space.
Our findings shed light on the ability of label-contrastive explanations to provide a more accurate and finer-grained interpretability of a model's decision.
arXiv Detail & Related papers (2021-03-02T00:36:45Z) - This is not the Texture you are looking for! Introducing Novel
Counterfactual Explanations for Non-Experts using Generative Adversarial
Learning [59.17685450892182]
counterfactual explanation systems try to enable a counterfactual reasoning by modifying the input image.
We present a novel approach to generate such counterfactual image explanations based on adversarial image-to-image translation techniques.
Our results show that our approach leads to significantly better results regarding mental models, explanation satisfaction, trust, emotions, and self-efficacy than two state-of-the art systems.
arXiv Detail & Related papers (2020-12-22T10:08:05Z) - Evaluating Explanations: How much do explanations from the teacher aid
students? [103.05037537415811]
We formalize the value of explanations using a student-teacher paradigm that measures the extent to which explanations improve student models in learning.
Unlike many prior proposals to evaluate explanations, our approach cannot be easily gamed, enabling principled, scalable, and automatic evaluation of attributions.
arXiv Detail & Related papers (2020-12-01T23:40:21Z) - A Diagnostic Study of Explainability Techniques for Text Classification [52.879658637466605]
We develop a list of diagnostic properties for evaluating existing explainability techniques.
We compare the saliency scores assigned by the explainability techniques with human annotations of salient input regions to find relations between a model's performance and the agreement of its rationales with human ones.
arXiv Detail & Related papers (2020-09-25T12:01:53Z) - A computational model implementing subjectivity with the 'Room Theory'.
The case of detecting Emotion from Text [68.8204255655161]
This work introduces a new method to consider subjectivity and general context dependency in text analysis.
By using similarity measure between words, we are able to extract the relative relevance of the elements in the benchmark.
This method could be applied to all the cases where evaluating subjectivity is relevant to understand the relative value or meaning of a text.
arXiv Detail & Related papers (2020-05-12T21:26:04Z) - Explainable Deep Classification Models for Domain Generalization [94.43131722655617]
Explanations are defined as regions of visual evidence upon which a deep classification network makes a decision.
Our training strategy enforces a periodic saliency-based feedback to encourage the model to focus on the image regions that directly correspond to the ground-truth object.
arXiv Detail & Related papers (2020-03-13T22:22:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.