BERT & Family Eat Word Salad: Experiments with Text Understanding
- URL: http://arxiv.org/abs/2101.03453v2
- Date: Wed, 17 Mar 2021 12:58:59 GMT
- Title: BERT & Family Eat Word Salad: Experiments with Text Understanding
- Authors: Ashim Gupta, Giorgi Kvernadze, Vivek Srikumar
- Abstract summary: We study the response of large models from the BERT family to incoherent inputs that should confuse any model that claims to understand natural language.
Experiments show that state-of-the-art models consistently fail to recognize them as ill-formed, and instead produce high confidence predictions on them.
We show that if models are explicitly trained to recognize invalid inputs, they can be robust to such attacks without a drop in performance.
- Score: 17.998891912502092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we study the response of large models from the BERT family to
incoherent inputs that should confuse any model that claims to understand
natural language. We define simple heuristics to construct such examples. Our
experiments show that state-of-the-art models consistently fail to recognize
them as ill-formed, and instead produce high confidence predictions on them. As
a consequence of this phenomenon, models trained on sentences with randomly
permuted word order perform close to state-of-the-art models. To alleviate
these issues, we show that if models are explicitly trained to recognize
invalid inputs, they can be robust to such attacks without a drop in
performance.
Related papers
- ALERT: Adapting Language Models to Reasoning Tasks [43.8679673685468]
ALERT is a benchmark and suite of analyses for assessing language models' reasoning ability.
ALERT provides a test bed to asses any language model on fine-grained reasoning skills.
We find that language models learn more reasoning skills during finetuning stage compared to pretraining state.
arXiv Detail & Related papers (2022-12-16T05:15:41Z) - Discovering Latent Knowledge in Language Models Without Supervision [72.95136739040676]
Existing techniques for training language models can be misaligned with the truth.
We propose directly finding latent knowledge inside the internal activations of a language model in a purely unsupervised way.
We show that despite using no supervision and no model outputs, our method can recover diverse knowledge represented in large language models.
arXiv Detail & Related papers (2022-12-07T18:17:56Z) - Does BERT really agree ? Fine-grained Analysis of Lexical Dependence on
a Syntactic Task [70.29624135819884]
We study the extent to which BERT is able to perform lexically-independent subject-verb number agreement (NA) on targeted syntactic templates.
Our results on nonce sentences suggest that the model generalizes well for simple templates, but fails to perform lexically-independent syntactic generalization when as little as one attractor is present.
arXiv Detail & Related papers (2022-04-14T11:33:15Z) - Interpreting Language Models with Contrastive Explanations [99.7035899290924]
Language models must consider various features to predict a token, such as its part of speech, number, tense, or semantics.
Existing explanation methods conflate evidence for all these features into a single explanation, which is less interpretable for human understanding.
We show that contrastive explanations are quantifiably better than non-contrastive explanations in verifying major grammatical phenomena.
arXiv Detail & Related papers (2022-02-21T18:32:24Z) - Explain, Edit, and Understand: Rethinking User Study Design for
Evaluating Model Explanations [97.91630330328815]
We conduct a crowdsourcing study, where participants interact with deception detection models that have been trained to distinguish between genuine and fake hotel reviews.
We observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
arXiv Detail & Related papers (2021-12-17T18:29:56Z) - Is My Model Using The Right Evidence? Systematic Probes for Examining
Evidence-Based Tabular Reasoning [26.168211982441875]
Neural models routinely report state-of-the-art performance across NLP tasks involving reasoning.
Our experiments demonstrate that a BERT-based model representative of today's state-of-the-art fails to properly reason on the following counts.
arXiv Detail & Related papers (2021-08-02T01:14:19Z) - What if This Modified That? Syntactic Interventions via Counterfactual
Embeddings [19.3614797257652]
Prior art aims to uncover meaningful properties within model representations via probes, but it is unclear how faithfully such probes portray information that the models actually use.
We propose a technique, inspired by causal analysis, for generating counterfactual embeddings within models.
In experiments testing our technique, we produce evidence that some BERT-based models use a tree-distance-like representation of syntax in downstream prediction tasks.
arXiv Detail & Related papers (2021-05-28T17:27:04Z) - Masked Language Modeling and the Distributional Hypothesis: Order Word
Matters Pre-training for Little [74.49773960145681]
A possible explanation for the impressive performance of masked language model (MLM)-training is that such models have learned to represent the syntactic structures prevalent in NLP pipelines.
In this paper, we propose a different explanation: pre-trains succeed on downstream tasks almost entirely due to their ability to model higher-order word co-occurrence statistics.
Our results show that purely distributional information largely explains the success of pre-training, and underscore the importance of curating challenging evaluation datasets that require deeper linguistic knowledge.
arXiv Detail & Related papers (2021-04-14T06:30:36Z) - Unnatural Language Inference [48.45003475966808]
We find that state-of-the-art NLI models, such as RoBERTa and BART, are invariant to, and sometimes even perform better on, examples with randomly reordered words.
Our findings call into question the idea that our natural language understanding models, and the tasks used for measuring their progress, genuinely require a human-like understanding of syntax.
arXiv Detail & Related papers (2020-12-30T20:40:48Z) - CausaLM: Causal Model Explanation Through Counterfactual Language Models [33.29636213961804]
CausaLM is a framework for producing causal model explanations using counterfactual language representation models.
We show that language representation models such as BERT can effectively learn a counterfactual representation for a given concept of interest.
A byproduct of our method is a language representation model that is unaffected by the tested concept.
arXiv Detail & Related papers (2020-05-27T15:06:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.