Symbolic Brittleness in Sequence Models: on Systematic Generalization in
Symbolic Mathematics
- URL: http://arxiv.org/abs/2109.13986v1
- Date: Tue, 28 Sep 2021 18:50:15 GMT
- Title: Symbolic Brittleness in Sequence Models: on Systematic Generalization in
Symbolic Mathematics
- Authors: Sean Welleck, Peter West, Jize Cao, Yejin Choi
- Abstract summary: We consider the problem of symbolic mathematical integration, as it requires generalizing systematically beyond the test set.
We develop a methodology for evaluating generalization that takes advantage of the problem domain's structure and access to a verifier.
We demonstrate challenges in achieving robustness, compositionality, and out-of-distribution generalization, through both carefully constructed manual test suites and a genetic algorithm.
- Score: 38.62999063710003
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural sequence models trained with maximum likelihood estimation have led to
breakthroughs in many tasks, where success is defined by the gap between
training and test performance. However, their ability to achieve stronger forms
of generalization remains unclear. We consider the problem of symbolic
mathematical integration, as it requires generalizing systematically beyond the
test set. We develop a methodology for evaluating generalization that takes
advantage of the problem domain's structure and access to a verifier. Despite
promising in-distribution performance of sequence-to-sequence models in this
domain, we demonstrate challenges in achieving robustness, compositionality,
and out-of-distribution generalization, through both carefully constructed
manual test suites and a genetic algorithm that automatically finds large
collections of failures in a controllable manner. Our investigation highlights
the difficulty of generalizing well with the predominant modeling and learning
approach, and the importance of evaluating beyond the test set, across
different aspects of generalization.
Related papers
- Topological Generalization Bounds for Discrete-Time Stochastic Optimization Algorithms [15.473123662393169]
Deep neural networks (DNNs) show remarkable generalization properties.
The source of these capabilities remains elusive, defying the established statistical learning theory.
Recent studies have revealed that properties of training trajectories can be indicative of generalization.
arXiv Detail & Related papers (2024-07-11T17:56:03Z) - Investigating the Role of Instruction Variety and Task Difficulty in Robotic Manipulation Tasks [50.75902473813379]
This work introduces a comprehensive evaluation framework that systematically examines the role of instructions and inputs in the generalisation abilities of such models.
The proposed framework uncovers the resilience of multimodal models to extreme instruction perturbations and their vulnerability to observational changes.
arXiv Detail & Related papers (2024-07-04T14:36:49Z) - A Comprehensive Augmentation Framework for Anomaly Detection [1.6114012813668932]
This paper analyzes crucial traits of simulated anomalies that contribute to the training of reconstructive networks.
We integrate this framework with a reconstruction-based approach and concurrently propose a split training strategy.
arXiv Detail & Related papers (2023-08-29T07:00:35Z) - A Hybrid System for Systematic Generalization in Simple Arithmetic
Problems [70.91780996370326]
We propose a hybrid system capable of solving arithmetic problems that require compositional and systematic reasoning over sequences of symbols.
We show that the proposed system can accurately solve nested arithmetical expressions even when trained only on a subset including the simplest cases.
arXiv Detail & Related papers (2023-06-29T18:35:41Z) - Understanding Generalization of Federated Learning via Stability:
Heterogeneity Matters [1.4502611532302039]
Generalization performance is a key metric in evaluating machine learning models when applied to real-world applications.
Generalization performance is a key metric in evaluating machine learning models when applied to real-world applications.
arXiv Detail & Related papers (2023-06-06T16:12:35Z) - Cross-functional Analysis of Generalisation in Behavioural Learning [4.0810783261728565]
We introduce BeLUGA, an analysis method for evaluating behavioural learning considering generalisation across dimensions of different levels.
An aggregate score measures generalisation to unseen functionalities (or overfitting)
arXiv Detail & Related papers (2023-05-22T11:54:19Z) - Neural Networks and the Chomsky Hierarchy [27.470857324448136]
We study whether insights from the theory of Chomsky can predict the limits of neural network generalization in practice.
We show negative results where even extensive amounts of data and training time never led to any non-trivial generalization.
Our results show that, for our subset of tasks, RNNs and Transformers fail to generalize on non-regular tasks, and only networks augmented with structured memory can successfully generalize on context-free and context-sensitive tasks.
arXiv Detail & Related papers (2022-07-05T15:06:11Z) - Amortized Inference for Causal Structure Learning [72.84105256353801]
Learning causal structure poses a search problem that typically involves evaluating structures using a score or independence test.
We train a variational inference model to predict the causal structure from observational/interventional data.
Our models exhibit robust generalization capabilities under substantial distribution shift.
arXiv Detail & Related papers (2022-05-25T17:37:08Z) - Generalization of Neural Combinatorial Solvers Through the Lens of
Adversarial Robustness [68.97830259849086]
Most datasets only capture a simpler subproblem and likely suffer from spurious features.
We study adversarial robustness - a local generalization property - to reveal hard, model-specific instances and spurious features.
Unlike in other applications, where perturbation models are designed around subjective notions of imperceptibility, our perturbation models are efficient and sound.
Surprisingly, with such perturbations, a sufficiently expressive neural solver does not suffer from the limitations of the accuracy-robustness trade-off common in supervised learning.
arXiv Detail & Related papers (2021-10-21T07:28:11Z) - Fractal Structure and Generalization Properties of Stochastic
Optimization Algorithms [71.62575565990502]
We prove that the generalization error of an optimization algorithm can be bounded on the complexity' of the fractal structure that underlies its generalization measure.
We further specialize our results to specific problems (e.g., linear/logistic regression, one hidden/layered neural networks) and algorithms.
arXiv Detail & Related papers (2021-06-09T08:05:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.