Investigating the Effect of Natural Language Explanations on
Out-of-Distribution Generalization in Few-shot NLI
- URL: http://arxiv.org/abs/2110.06223v1
- Date: Tue, 12 Oct 2021 18:00:02 GMT
- Title: Investigating the Effect of Natural Language Explanations on
Out-of-Distribution Generalization in Few-shot NLI
- Authors: Yangqiaoyu Zhou, Chenhao Tan
- Abstract summary: We formulate a few-shot learning setup and examine the effects of natural language explanations on OOD generalization.
We leverage the templates in the HANS dataset and construct templated natural language explanations for each template.
We show that generated explanations show competitive BLEU scores against groundtruth explanations, but they fail to improve prediction performance.
- Score: 11.44224857047629
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although neural models have shown strong performance in datasets such as
SNLI, they lack the ability to generalize out-of-distribution (OOD). In this
work, we formulate a few-shot learning setup and examine the effects of natural
language explanations on OOD generalization. We leverage the templates in the
HANS dataset and construct templated natural language explanations for each
template. Although generated explanations show competitive BLEU scores against
groundtruth explanations, they fail to improve prediction performance. We
further show that generated explanations often hallucinate information and miss
key elements that indicate the label.
Related papers
- Properties and Challenges of LLM-Generated Explanations [3.257973235065581]
We study the self-rationalising capabilities of large language models (LLMs)
We find that generated explanations show selectivity and contain illustrative elements, but less frequently are subjective or misleading.
In particular, we outline positive and negative implications depending on the goals and user groups of the self-rationalising system.
arXiv Detail & Related papers (2024-02-16T09:37:54Z) - FLamE: Few-shot Learning from Natural Language Explanations [12.496665033682202]
We present FLamE, a framework for learning from natural language explanations.
Experiments on natural language inference demonstrate effectiveness over strong baselines.
Human evaluation surprisingly reveals that the majority of generated explanations does not adequately justify classification decisions.
arXiv Detail & Related papers (2023-06-13T18:01:46Z) - Simple Linguistic Inferences of Large Language Models (LLMs): Blind Spots and Blinds [59.71218039095155]
We evaluate language understanding capacities on simple inference tasks that most humans find trivial.
We target (i) grammatically-specified entailments, (ii) premises with evidential adverbs of uncertainty, and (iii) monotonicity entailments.
The models exhibit moderate to low performance on these evaluation sets.
arXiv Detail & Related papers (2023-05-24T06:41:09Z) - FOLIO: Natural Language Reasoning with First-Order Logic [147.50480350846726]
We present FOLIO, a human-annotated, logically complex and diverse dataset for reasoning in natural language (NL)
FOLIO consists of 1,430 examples (unique conclusions), each paired with one of 487 sets of premises used to deductively reason for the validity of each conclusion.
For both NL reasoning and NL-FOL translation, we benchmark multiple state-of-the-art language models.
arXiv Detail & Related papers (2022-09-02T06:50:11Z) - Does BERT really agree ? Fine-grained Analysis of Lexical Dependence on
a Syntactic Task [70.29624135819884]
We study the extent to which BERT is able to perform lexically-independent subject-verb number agreement (NA) on targeted syntactic templates.
Our results on nonce sentences suggest that the model generalizes well for simple templates, but fails to perform lexically-independent syntactic generalization when as little as one attractor is present.
arXiv Detail & Related papers (2022-04-14T11:33:15Z) - Masked Language Modeling and the Distributional Hypothesis: Order Word
Matters Pre-training for Little [74.49773960145681]
A possible explanation for the impressive performance of masked language model (MLM)-training is that such models have learned to represent the syntactic structures prevalent in NLP pipelines.
In this paper, we propose a different explanation: pre-trains succeed on downstream tasks almost entirely due to their ability to model higher-order word co-occurrence statistics.
Our results show that purely distributional information largely explains the success of pre-training, and underscore the importance of curating challenging evaluation datasets that require deeper linguistic knowledge.
arXiv Detail & Related papers (2021-04-14T06:30:36Z) - Towards Interpretable Natural Language Understanding with Explanations
as Latent Variables [146.83882632854485]
We develop a framework for interpretable natural language understanding that requires only a small set of human annotated explanations for training.
Our framework treats natural language explanations as latent variables that model the underlying reasoning process of a neural model.
arXiv Detail & Related papers (2020-10-24T02:05:56Z) - Leakage-Adjusted Simulatability: Can Models Generate Non-Trivial
Explanations of Their Behavior in Natural Language? [86.60613602337246]
We introduce a leakage-adjusted simulatability (LAS) metric for evaluating NL explanations.
LAS measures how well explanations help an observer predict a model's output, while controlling for how explanations can directly leak the output.
We frame explanation generation as a multi-agent game and optimize explanations for simulatability while penalizing label leakage.
arXiv Detail & Related papers (2020-10-08T16:59:07Z) - NILE : Natural Language Inference with Faithful Natural Language
Explanations [10.074153632701952]
We propose Natural-language Inference over Label-specific Explanations (NILE)
NILE is a novel NLI method which utilizes auto-generated label-specific explanations to produce labels along with its faithful explanation.
We discuss the faithfulness of NILE's explanations in terms of sensitivity of the decisions to the corresponding explanations.
arXiv Detail & Related papers (2020-05-25T13:56:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.