Estimating the Causal Effects of Natural Logic Features in Neural NLI
Models
- URL: http://arxiv.org/abs/2305.08572v1
- Date: Mon, 15 May 2023 12:01:09 GMT
- Title: Estimating the Causal Effects of Natural Logic Features in Neural NLI
Models
- Authors: Julia Rozanova, Marco Valentino, Andre Freitas
- Abstract summary: We zone in on specific patterns of reasoning with enough structure and regularity to be able to identify and quantify systematic reasoning failures in widely-used models.
We apply causal effect estimation strategies to measure the effect of context interventions.
Following related work on causal analysis of NLP models in different settings, we adapt the methodology for the NLI task to construct comparative model profiles.
- Score: 2.363388546004777
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Rigorous evaluation of the causal effects of semantic features on language
model predictions can be hard to achieve for natural language reasoning
problems. However, this is such a desirable form of analysis from both an
interpretability and model evaluation perspective, that it is valuable
to zone in on specific patterns of reasoning with enough structure and
regularity
to be able to identify and quantify systematic reasoning failures in
widely-used models. In this vein, we pick a portion of the NLI task for which
an explicit causal diagram can be systematically constructed: in particular,
the case where across two sentences (the premise and hypothesis), two related
words/terms occur in a shared context.
In this work, we apply causal effect estimation strategies to measure the
effect of context interventions
(whose effect on the entailment label is mediated by the semantic
monotonicity characteristic) and interventions on the inserted
word-pair (whose effect on the entailment label is mediated by the relation
between these words.).
Following related work on causal analysis of NLP models in different
settings, we
adapt the methodology for the NLI task to construct comparative model
profiles in terms of robustness to irrelevant changes and sensitivity to
impactful changes.
Related papers
- Estimating the Causal Effects of Natural Logic Features in Transformer-Based NLI Models [16.328341121232484]
We apply causal effect estimation strategies to measure the effect of context interventions.
We investigate robustness to irrelevant changes and sensitivity to impactful changes of Transformers.
arXiv Detail & Related papers (2024-04-03T10:22:35Z) - Semantic Sensitivities and Inconsistent Predictions: Measuring the
Fragility of NLI Models [44.56781176879151]
State-of-the-art Natural Language Inference (NLI) models are sensitive towards minor semantics preserving surface-form variations.
We show that semantic sensitivity causes performance degradations of $12.92%$ and $23.71%$ average over $textbfin-$ and $textbfout-of-$ domain settings.
arXiv Detail & Related papers (2024-01-25T14:47:05Z) - Identifiable Latent Polynomial Causal Models Through the Lens of Change [82.14087963690561]
Causal representation learning aims to unveil latent high-level causal representations from observed low-level data.
One of its primary tasks is to provide reliable assurance of identifying these latent causal models, known as identifiability.
arXiv Detail & Related papers (2023-10-24T07:46:10Z) - Inducing Causal Structure for Abstractive Text Summarization [76.1000380429553]
We introduce a Structural Causal Model (SCM) to induce the underlying causal structure of the summarization data.
We propose a Causality Inspired Sequence-to-Sequence model (CI-Seq2Seq) to learn the causal representations that can mimic the causal factors.
Experimental results on two widely used text summarization datasets demonstrate the advantages of our approach.
arXiv Detail & Related papers (2023-08-24T16:06:36Z) - Interventional Probing in High Dimensions: An NLI Case Study [2.1028463367241033]
Probing strategies have been shown to detect semantic features intermediate to the "natural logic" fragment of the Natural Language Inference task (NLI)
In this work, we carry out new and existing representation-level interventions to investigate the effect of these semantic features on NLI classification.
arXiv Detail & Related papers (2023-04-20T14:34:31Z) - CausalDialogue: Modeling Utterance-level Causality in Conversations [83.03604651485327]
We have compiled and expanded upon a new dataset called CausalDialogue through crowd-sourcing.
This dataset includes multiple cause-effect pairs within a directed acyclic graph (DAG) structure.
We propose a causality-enhanced method called Exponential Average Treatment Effect (ExMATE) to enhance the impact of causality at the utterance level in training neural conversation models.
arXiv Detail & Related papers (2022-12-20T18:31:50Z) - Estimation of Bivariate Structural Causal Models by Variational Gaussian
Process Regression Under Likelihoods Parametrised by Normalising Flows [74.85071867225533]
Causal mechanisms can be described by structural causal models.
One major drawback of state-of-the-art artificial intelligence is its lack of explainability.
arXiv Detail & Related papers (2021-09-06T14:52:58Z) - Did the Cat Drink the Coffee? Challenging Transformers with Generalized
Event Knowledge [59.22170796793179]
Transformers Language Models (TLMs) were tested on a benchmark for the textitdynamic estimation of thematic fit
Our results show that TLMs can reach performances that are comparable to those achieved by SDM.
However, additional analysis consistently suggests that TLMs do not capture important aspects of event knowledge.
arXiv Detail & Related papers (2021-07-22T20:52:26Z) - Learning Causal Semantic Representation for Out-of-Distribution
Prediction [125.38836464226092]
We propose a Causal Semantic Generative model (CSG) based on a causal reasoning so that the two factors are modeled separately.
We show that CSG can identify the semantic factor by fitting training data, and this semantic-identification guarantees the boundedness of OOD generalization error.
arXiv Detail & Related papers (2020-11-03T13:16:05Z) - Estimating Causal Effects with the Neural Autoregressive Density
Estimator [6.59529078336196]
We use neural autoregressive density estimators to estimate causal effects within the Pearl's do-calculus framework.
We show that the approach can retrieve causal effects from non-linear systems without explicitly modeling the interactions between the variables.
arXiv Detail & Related papers (2020-08-17T13:12:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.