On Interactive Explanations as Non-Monotonic Reasoning
- URL: http://arxiv.org/abs/2208.00316v1
- Date: Sat, 30 Jul 2022 22:08:35 GMT
- Title: On Interactive Explanations as Non-Monotonic Reasoning
- Authors: Guilherme Paulino-Passos and Francesca Toni
- Abstract summary: We treat explanations as objects that can be subject to reasoning.
We present a formal model of the interactive scenario between user and system.
This allows: 1) to solve some considered inconsistencies in explanation, such as via a specificity relation; 2) to consider properties from the non-monotonic reasoning literature and discuss their desirability.
- Score: 10.616061367794385
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work shows issues of consistency with explanations, with methods
generating local explanations that seem reasonable instance-wise, but that are
inconsistent across instances. This suggests not only that instance-wise
explanations can be unreliable, but mainly that, when interacting with a system
via multiple inputs, a user may actually lose confidence in the system. To
better analyse this issue, in this work we treat explanations as objects that
can be subject to reasoning and present a formal model of the interactive
scenario between user and system, via sequences of inputs, outputs, and
explanations. We argue that explanations can be thought of as committing to
some model behaviour (even if only prima facie), suggesting a form of
entailment, which, we argue, should be thought of as non-monotonic. This
allows: 1) to solve some considered inconsistencies in explanation, such as via
a specificity relation; 2) to consider properties from the non-monotonic
reasoning literature and discuss their desirability, gaining more insight on
the interactive explanation scenario.
Related papers
- Explaining Explanations in Probabilistic Logic Programming [0.0]
In most approaches, the system is considered a black box, making it difficult to generate appropriate explanations.
We consider a setting where models are transparent: probabilistic logic programming (PLP), a paradigm that combines logic programming for knowledge representation and probability to model uncertainty.
We present in this paper an approach to explaining explanations which is based on defining a new query-driven inference mechanism for PLP where proofs are labeled with "choice expressions", a compact and easy to manipulate representation for sets of choices.
arXiv Detail & Related papers (2024-01-30T14:27:37Z) - Abductive Commonsense Reasoning Exploiting Mutually Exclusive
Explanations [118.0818807474809]
Abductive reasoning aims to find plausible explanations for an event.
Existing approaches for abductive reasoning in natural language processing often rely on manually generated annotations for supervision.
This work proposes an approach for abductive commonsense reasoning that exploits the fact that only a subset of explanations is correct for a given context.
arXiv Detail & Related papers (2023-05-24T01:35:10Z) - HOP, UNION, GENERATE: Explainable Multi-hop Reasoning without Rationale
Supervision [118.0818807474809]
This work proposes a principled, probabilistic approach for training explainable multi-hop QA systems without rationale supervision.
Our approach performs multi-hop reasoning by explicitly modeling rationales as sets, enabling the model to capture interactions between documents and sentences within a document.
arXiv Detail & Related papers (2023-05-23T16:53:49Z) - Explanation Selection Using Unlabeled Data for Chain-of-Thought
Prompting [80.9896041501715]
Explanations that have not been "tuned" for a task, such as off-the-shelf explanations written by nonexperts, may lead to mediocre performance.
This paper tackles the problem of how to optimize explanation-infused prompts in a blackbox fashion.
arXiv Detail & Related papers (2023-02-09T18:02:34Z) - MetaLogic: Logical Reasoning Explanations with Fine-Grained Structure [129.8481568648651]
We propose a benchmark to investigate models' logical reasoning capabilities in complex real-life scenarios.
Based on the multi-hop chain of reasoning, the explanation form includes three main components.
We evaluate the current best models' performance on this new explanation form.
arXiv Detail & Related papers (2022-10-22T16:01:13Z) - Logical Satisfiability of Counterfactuals for Faithful Explanations in
NLI [60.142926537264714]
We introduce the methodology of Faithfulness-through-Counterfactuals.
It generates a counterfactual hypothesis based on the logical predicates expressed in the explanation.
It then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic.
arXiv Detail & Related papers (2022-05-25T03:40:59Z) - Contrastive Explanations for Model Interpretability [77.92370750072831]
We propose a methodology to produce contrastive explanations for classification models.
Our method is based on projecting model representation to a latent space.
Our findings shed light on the ability of label-contrastive explanations to provide a more accurate and finer-grained interpretability of a model's decision.
arXiv Detail & Related papers (2021-03-02T00:36:45Z) - Semantics and explanation: why counterfactual explanations produce
adversarial examples in deep neural networks [15.102346715690759]
Recent papers in explainable AI have made a compelling case for counterfactual modes of explanation.
While counterfactual explanations appear to be extremely effective in some instances, they are formally equivalent to adversarial examples.
This presents an apparent paradox for explainability researchers: if these two procedures are formally equivalent, what accounts for the explanatory divide apparent between counterfactual explanations and adversarial examples?
We resolve this paradox by placing emphasis back on the semantics of counterfactual expressions.
arXiv Detail & Related papers (2020-12-18T07:04:04Z) - LIMEtree: Interactively Customisable Explanations Based on Local
Surrogate Multi-output Regression Trees [21.58324172085553]
We introduce a model-agnostic and post-hoc local explainability technique for black-box predictions called LIMEtree.
We validate our algorithm on a deep neural network trained for object detection in images and compare it against Local Interpretable Model-agnostic Explanations (LIME)
Our method comes with local fidelity guarantees and can produce a range of diverse explanation types.
arXiv Detail & Related papers (2020-05-04T12:31:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.