Towards Coinductive Models for Natural Language Understanding. Bringing
together Deep Learning and Deep Semantics
- URL: http://arxiv.org/abs/2012.05715v1
- Date: Wed, 9 Dec 2020 03:10:36 GMT
- Title: Towards Coinductive Models for Natural Language Understanding. Bringing
together Deep Learning and Deep Semantics
- Authors: Wlodek W. Zadrozny
- Abstract summary: Coinduction has been successfully used in the design of operating systems and programming languages.
It has been present in text mining, machine translation, and in some attempts to model intensionality and modalities.
This article shows several examples of the joint appearance of induction and coinduction in natural language processing.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This article contains a proposal to add coinduction to the computational
apparatus of natural language understanding. This, we argue, will provide a
basis for more realistic, computationally sound, and scalable models of natural
language dialogue, syntax and semantics. Given that the bottom up, inductively
constructed, semantic and syntactic structures are brittle, and seemingly
incapable of adequately representing the meaning of longer sentences or
realistic dialogues, natural language understanding is in need of a new
foundation. Coinduction, which uses top down constraints, has been successfully
used in the design of operating systems and programming languages. Moreover,
implicitly it has been present in text mining, machine translation, and in some
attempts to model intensionality and modalities, which provides evidence that
it works. This article shows high level formalizations of some of such uses.
Since coinduction and induction can coexist, they can provide a common
language and a conceptual model for research in natural language understanding.
In particular, such an opportunity seems to be emerging in research on
compositionality. This article shows several examples of the joint appearance
of induction and coinduction in natural language processing. We argue that the
known individual limitations of induction and coinduction can be overcome in
empirical settings by a combination of the the two methods. We see an open
problem in providing a theory of their joint use.
Related papers
- A Complexity-Based Theory of Compositionality [53.025566128892066]
In AI, compositional representations can enable a powerful form of out-of-distribution generalization.
Here, we propose a formal definition of compositionality that accounts for and extends our intuitions about compositionality.
The definition is conceptually simple, quantitative, grounded in algorithmic information theory, and applicable to any representation.
arXiv Detail & Related papers (2024-10-18T18:37:27Z) - From Word Models to World Models: Translating from Natural Language to
the Probabilistic Language of Thought [124.40905824051079]
We propose rational meaning construction, a computational framework for language-informed thinking.
We frame linguistic meaning as a context-sensitive mapping from natural language into a probabilistic language of thought.
We show that LLMs can generate context-sensitive translations that capture pragmatically-appropriate linguistic meanings.
We extend our framework to integrate cognitively-motivated symbolic modules.
arXiv Detail & Related papers (2023-06-22T05:14:00Z) - Large Language Models are In-Context Semantic Reasoners rather than
Symbolic Reasoners [75.85554779782048]
Large Language Models (LLMs) have excited the natural language and machine learning community over recent years.
Despite of numerous successful applications, the underlying mechanism of such in-context capabilities still remains unclear.
In this work, we hypothesize that the learned textitsemantics of language tokens do the most heavy lifting during the reasoning process.
arXiv Detail & Related papers (2023-05-24T07:33:34Z) - Language Models as Inductive Reasoners [125.99461874008703]
We propose a new paradigm (task) for inductive reasoning, which is to induce natural language rules from natural language facts.
We create a dataset termed DEER containing 1.2k rule-fact pairs for the task, where rules and facts are written in natural language.
We provide the first and comprehensive analysis of how well pretrained language models can induce natural language rules from natural language facts.
arXiv Detail & Related papers (2022-12-21T11:12:14Z) - Transparency Helps Reveal When Language Models Learn Meaning [71.96920839263457]
Our systematic experiments with synthetic data reveal that, with languages where all expressions have context-independent denotations, both autoregressive and masked language models learn to emulate semantic relations between expressions.
Turning to natural language, our experiments with a specific phenomenon -- referential opacity -- add to the growing body of evidence that current language models do not well-represent natural language semantics.
arXiv Detail & Related papers (2022-10-14T02:35:19Z) - A Model of Anaphoric Ambiguities using Sheaf Theoretic Quantum-like
Contextuality and BERT [0.0]
We construct a schema for anaphoric ambiguities that exhibits quantum-like contextuality.
We then take advantage of the neural word embedding engine BERT to instantiate the schema to natural language examples.
Our hope is that these examples will pave the way for future research and for finding ways to extend applications of quantum computing to natural language processing.
arXiv Detail & Related papers (2022-08-11T09:31:15Z) - Target Languages (vs. Inductive Biases) for Learning to Act and Plan [13.820550902006078]
I articulate a different learning approach where representations do not emerge from biases in a neural architecture but are learned over a given target language with a known semantics.
The goals of the paper and talk are to make these ideas explicit, to place them in a broader context where the design of the target language is crucial, and to illustrate them in the context of learning to act and plan.
arXiv Detail & Related papers (2021-09-15T10:24:13Z) - Compositional Processing Emerges in Neural Networks Solving Math
Problems [100.80518350845668]
Recent progress in artificial neural networks has shown that when large models are trained on enough linguistic data, grammatical structure emerges in their representations.
We extend this work to the domain of mathematical reasoning, where it is possible to formulate precise hypotheses about how meanings should be composed.
Our work shows that neural networks are not only able to infer something about the structured relationships implicit in their training data, but can also deploy this knowledge to guide the composition of individual meanings into composite wholes.
arXiv Detail & Related papers (2021-05-19T07:24:42Z) - Modelling Compositionality and Structure Dependence in Natural Language [0.12183405753834563]
Drawing on linguistics and set theory, a formalisation of these ideas is presented in the first half of this thesis.
We see how cognitive systems that process language need to have certain functional constraints.
Using the advances of word embedding techniques, a model of relational learning is simulated.
arXiv Detail & Related papers (2020-11-22T17:28:50Z) - Universal linguistic inductive biases via meta-learning [36.43388942327124]
It is unclear which inductive biases can explain observed patterns in language acquisition.
We introduce a framework for giving linguistic inductive biases to a neural network model.
We demonstrate this framework with a case study based on syllable structure.
arXiv Detail & Related papers (2020-06-29T19:15:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.