Probing Linguistic Systematicity
- URL: http://arxiv.org/abs/2005.04315v2
- Date: Tue, 25 Aug 2020 04:18:42 GMT
- Title: Probing Linguistic Systematicity
- Authors: Emily Goodwin and Koustuv Sinha and Timothy J. O'Donnell
- Abstract summary: There is accumulating evidence that neural models often generalize non-systematically.
We identify ways in which network architectures can generalize non-systematically, and discuss why such forms of generalization may be unsatisfying.
- Score: 11.690179162556353
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, there has been much interest in the question of whether deep
natural language understanding models exhibit systematicity; generalizing such
that units like words make consistent contributions to the meaning of the
sentences in which they appear. There is accumulating evidence that neural
models often generalize non-systematically. We examined the notion of
systematicity from a linguistic perspective, defining a set of probes and a set
of metrics to measure systematic behaviour. We also identified ways in which
network architectures can generalize non-systematically, and discuss why such
forms of generalization may be unsatisfying. As a case study, we performed a
series of experiments in the setting of natural language inference (NLI),
demonstrating that some NLU systems achieve high overall performance despite
being non-systematic.
Related papers
- Architecture of a Cortex Inspired Hierarchical Event Recaller [0.0]
This paper proposes a new approach to Machine Learning (ML) that focuses on unsupervised continuous context-dependent learning of complex patterns.
A synthetic structure capable of identifying and predicting complex temporal series will be defined and experimentally tested.
As a proof of concept, the proposed system is shown to be able to learn, identify and predict a remarkably complex temporal series such as human speech, with no prior knowledge.
arXiv Detail & Related papers (2024-05-03T09:36:16Z) - Generative Models as a Complex Systems Science: How can we make sense of
large language model behavior? [75.79305790453654]
Coaxing out desired behavior from pretrained models, while avoiding undesirable ones, has redefined NLP.
We argue for a systematic effort to decompose language model behavior into categories that explain cross-task performance.
arXiv Detail & Related papers (2023-07-31T22:58:41Z) - A Hybrid System for Systematic Generalization in Simple Arithmetic
Problems [70.91780996370326]
We propose a hybrid system capable of solving arithmetic problems that require compositional and systematic reasoning over sequences of symbols.
We show that the proposed system can accurately solve nested arithmetical expressions even when trained only on a subset including the simplest cases.
arXiv Detail & Related papers (2023-06-29T18:35:41Z) - Neural-Symbolic Recursive Machine for Systematic Generalization [113.22455566135757]
We introduce the Neural-Symbolic Recursive Machine (NSR), whose core is a Grounded Symbol System (GSS)
NSR integrates neural perception, syntactic parsing, and semantic reasoning.
We evaluate NSR's efficacy across four challenging benchmarks designed to probe systematic generalization capabilities.
arXiv Detail & Related papers (2022-10-04T13:27:38Z) - Compositional Semantics and Inference System for Temporal Order based on
Japanese CCG [9.683269364766426]
We present a logic-based Natural Language Inference system that considers temporal order in Japanese.
Our system performs inference involving temporal order by using axioms for temporal relations and automated theorem provers.
We show that our system outperforms previous logic-based systems as well as current deep learning-based models.
arXiv Detail & Related papers (2022-04-20T06:21:21Z) - SyGNS: A Systematic Generalization Testbed Based on Natural Language
Semantics [39.845425535943534]
We propose a Systematic Generalization testbed based on Natural language Semantics (SyGNS)
We test whether neural networks can systematically parse sentences involving novel combinations of logical expressions such as quantifiers and negation.
Experiments show that Transformer and GRU models can generalize to unseen combinations of quantifiers, negations, and modifier that are similar to given training instances in form, but not to the others.
arXiv Detail & Related papers (2021-06-02T11:24:41Z) - Discrete representations in neural models of spoken language [56.29049879393466]
We compare the merits of four commonly used metrics in the context of weakly supervised models of spoken language.
We find that the different evaluation metrics can give inconsistent results.
arXiv Detail & Related papers (2021-05-12T11:02:02Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Curious Case of Language Generation Evaluation Metrics: A Cautionary
Tale [52.663117551150954]
A few popular metrics remain as the de facto metrics to evaluate tasks such as image captioning and machine translation.
This is partly due to ease of use, and partly because researchers expect to see them and know how to interpret them.
In this paper, we urge the community for more careful consideration of how they automatically evaluate their models.
arXiv Detail & Related papers (2020-10-26T13:57:20Z) - Systematic Generalization on gSCAN with Language Conditioned Embedding [19.39687991647301]
Systematic Generalization refers to a learning algorithm's ability to extrapolate learned behavior to unseen situations.
We propose a novel method that learns objects' contextualized embeddings with dynamic message passing conditioned on the input natural language.
arXiv Detail & Related papers (2020-09-11T17:35:05Z) - Do Neural Models Learn Systematicity of Monotonicity Inference in
Natural Language? [41.649440404203595]
We introduce a method for evaluating whether neural models can learn systematicity of monotonicity inference in natural language.
We consider four aspects of monotonicity inferences and test whether the models can systematically interpret lexical and logical phenomena on different training/test splits.
arXiv Detail & Related papers (2020-04-30T14:48:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.