Case-Based Abductive Natural Language Inference
- URL: http://arxiv.org/abs/2009.14539v4
- Date: Sat, 10 Sep 2022 09:07:57 GMT
- Title: Case-Based Abductive Natural Language Inference
- Authors: Marco Valentino, Mokanarangan Thayaparan, Andr\'e Freitas
- Abstract summary: Case-Based Abductive Natural Language Inference (CB-ANLI)
Case-Based Abductive Natural Language Inference (CB-ANLI)
- Score: 4.726777092009554
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most of the contemporary approaches for multi-hop Natural Language Inference
(NLI) construct explanations considering each test case in isolation. However,
this paradigm is known to suffer from semantic drift, a phenomenon that causes
the construction of spurious explanations leading to wrong conclusions. In
contrast, this paper proposes an abductive framework for multi-hop NLI
exploring the retrieve-reuse-refine paradigm in Case-Based Reasoning (CBR).
Specifically, we present Case-Based Abductive Natural Language Inference
(CB-ANLI), a model that addresses unseen inference problems by analogical
transfer of prior explanations from similar examples. We empirically evaluate
the abductive framework on commonsense and scientific question answering tasks,
demonstrating that CB-ANLI can be effectively integrated with sparse and dense
pre-trained encoders to improve multi-hop inference, or adopted as an evidence
retriever for Transformers. Moreover, an empirical analysis of semantic drift
reveals that the CBR paradigm boosts the quality of the most challenging
explanations, a feature that has a direct impact on robustness and accuracy in
downstream inference tasks.
Related papers
- Explaining Text Similarity in Transformer Models [52.571158418102584]
Recent advances in explainable AI have made it possible to mitigate limitations by leveraging improved explanations for Transformers.
We use BiLRP, an extension developed for computing second-order explanations in bilinear similarity models, to investigate which feature interactions drive similarity in NLP models.
Our findings contribute to a deeper understanding of different semantic similarity tasks and models, highlighting how novel explainable AI methods enable in-depth analyses and corpus-level insights.
arXiv Detail & Related papers (2024-05-10T17:11:31Z) - Advancing Counterfactual Inference through Nonlinear Quantile Regression [77.28323341329461]
We propose a framework for efficient and effective counterfactual inference implemented with neural networks.
The proposed approach enhances the capacity to generalize estimated counterfactual outcomes to unseen data.
Empirical results conducted on multiple datasets offer compelling support for our theoretical assertions.
arXiv Detail & Related papers (2023-06-09T08:30:51Z) - Learning a Structural Causal Model for Intuition Reasoning in
Conversation [20.243323155177766]
Reasoning, a crucial aspect of NLP research, has not been adequately addressed by prevailing models.
We develop a conversation cognitive model ( CCM) that explains how each utterance receives and activates channels of information.
By leveraging variational inference, it explores substitutes for implicit causes, addresses the issue of their unobservability, and reconstructs the causal representations of utterances through the evidence lower bounds.
arXiv Detail & Related papers (2023-05-28T13:54:09Z) - Counterfactuals of Counterfactuals: a back-translation-inspired approach
to analyse counterfactual editors [3.4253416336476246]
We focus on the analysis of counterfactual, contrastive explanations.
We propose a new back translation-inspired evaluation methodology.
We show that by iteratively feeding the counterfactual to the explainer we can obtain valuable insights into the behaviour of both the predictor and the explainer models.
arXiv Detail & Related papers (2023-05-26T16:04:28Z) - Abductive Commonsense Reasoning Exploiting Mutually Exclusive
Explanations [118.0818807474809]
Abductive reasoning aims to find plausible explanations for an event.
Existing approaches for abductive reasoning in natural language processing often rely on manually generated annotations for supervision.
This work proposes an approach for abductive commonsense reasoning that exploits the fact that only a subset of explanations is correct for a given context.
arXiv Detail & Related papers (2023-05-24T01:35:10Z) - Logical Satisfiability of Counterfactuals for Faithful Explanations in
NLI [60.142926537264714]
We introduce the methodology of Faithfulness-through-Counterfactuals.
It generates a counterfactual hypothesis based on the logical predicates expressed in the explanation.
It then evaluates if the model's prediction on the counterfactual is consistent with that expressed logic.
arXiv Detail & Related papers (2022-05-25T03:40:59Z) - Probing as Quantifying the Inductive Bias of Pre-trained Representations [99.93552997506438]
We present a novel framework for probing where the goal is to evaluate the inductive bias of representations for a particular task.
We apply our framework to a series of token-, arc-, and sentence-level tasks.
arXiv Detail & Related papers (2021-10-15T22:01:16Z) - Hybrid Autoregressive Solver for Scalable Abductive Natural Language
Inference [2.867517731896504]
We propose a hybrid abductive solver that autoregressively combines a dense bi-encoder with a sparse model of explanatory power.
Our experiments demonstrate that the proposed framework can achieve performance comparable with the state-of-the-art cross-encoder.
arXiv Detail & Related papers (2021-07-25T19:29:53Z) - Structural Causal Models Reveal Confounder Bias in Linear Program
Modelling [26.173103098250678]
We investigate the question of whether the phenomenon might be more general in nature, that is, adversarial-style attacks outside classical classification tasks.
Specifically, we consider the base class of Linear Programs (LPs)
We show the direct influence of the Structural Causal Model (SCM) onto the subsequent LP optimization, which ultimately exposes a notion of confounding in LPs.
arXiv Detail & Related papers (2021-05-26T17:19:22Z) - Influence-Driven Explanations for Bayesian Network Classifiers [16.708069984516964]
We focus on explanations for discrete network Bayesian classifiers (BCs)
We propose influence-driven explanations (IDXs) for BCs that include intermediate variables in explanations, rather than just the input and output variables as is standard practice.
arXiv Detail & Related papers (2020-12-10T16:00:51Z) - Learning Causal Semantic Representation for Out-of-Distribution
Prediction [125.38836464226092]
We propose a Causal Semantic Generative model (CSG) based on a causal reasoning so that the two factors are modeled separately.
We show that CSG can identify the semantic factor by fitting training data, and this semantic-identification guarantees the boundedness of OOD generalization error.
arXiv Detail & Related papers (2020-11-03T13:16:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.