Traduction des Grammaires Cat\'egorielles de Lambek dans les Grammaires
Cat\'egorielles Abstraites
- URL: http://arxiv.org/abs/2002.00725v1
- Date: Thu, 23 Jan 2020 18:23:03 GMT
- Title: Traduction des Grammaires Cat\'egorielles de Lambek dans les Grammaires
Cat\'egorielles Abstraites
- Authors: Valentin D. Richard
- Abstract summary: This internship report is to demonstrate that every Lambek Grammar can be, not entirely but efficiently, expressed in Abstract Categorial Grammars (ACG)
The main idea is to transform the type rewriting system of LGs into that of Context-Free Grammars (CFG) by erasing introduction and elimination rules and generating enough axioms so that the cut rule suffices.
Although the underlying algorithm was not fully implemented, this proof provides another argument in favour of the relevance of ACGs in Natural Language Processing.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lambek Grammars (LG) are a computational modelling of natural language, based
on non-commutative compositional types. It has been widely studied, especially
for languages where the syntax plays a major role (like English). The goal of
this internship report is to demonstrate that every Lambek Grammar can be, not
entirely but efficiently, expressed in Abstract Categorial Grammars (ACG). The
latter is a novel modelling based on higher-order signature homomorphisms
(using $\lambda$-calculus), aiming at uniting the currently used models. The
main idea is to transform the type rewriting system of LGs into that of
Context-Free Grammars (CFG) by erasing introduction and elimination rules and
generating enough axioms so that the cut rule suffices. This iterative approach
preserves the derivations and enables us to stop the possible infinite
generative process at any step. Although the underlying algorithm was not fully
implemented, this proof provides another argument in favour of the relevance of
ACGs in Natural Language Processing.
Related papers
- Authorship Verification based on the Likelihood Ratio of Grammar Models [0.8749675983608172]
Authorship Verification (AV) is the process of analyzing a set of documents to determine whether they were written by a specific author.
We propose a method relying on calculating a quantity we call $lambda_G$ (LambdaG)
Despite not needing large amounts of data for training, LambdaG still outperforms other established AV methods with higher computational complexity.
arXiv Detail & Related papers (2024-03-13T12:25:47Z) - Lexinvariant Language Models [84.2829117441298]
Token embeddings, a mapping from discrete lexical symbols to continuous vectors, are at the heart of any language model (LM)
We study textitlexinvariantlanguage models that are invariant to lexical symbols and therefore do not need fixed token embeddings in practice.
We show that a lexinvariant LM can attain perplexity comparable to that of a standard language model, given a sufficiently long context.
arXiv Detail & Related papers (2023-05-24T19:10:46Z) - Physics of Language Models: Part 1, Learning Hierarchical Language Structures [51.68385617116854]
Transformer-based language models are effective but complex, and understanding their inner workings is a significant challenge.
We introduce a family of synthetic CFGs that produce hierarchical rules, capable of generating lengthy sentences.
We demonstrate that generative models like GPT can accurately learn this CFG language and generate sentences based on it.
arXiv Detail & Related papers (2023-05-23T04:28:16Z) - Language Models of Code are Few-Shot Commonsense Learners [106.1531522893209]
Given a natural language input, the goal is to generate a graph such as an event -- or a reasoning-graph.
Existing approaches serialize the output graph as a flat list of nodes and edges.
We show that when we instead frame structured commonsense reasoning tasks as code generation tasks, pre-trained LMs of code are better structured commonsense reasoners than LMs of natural language.
arXiv Detail & Related papers (2022-10-13T16:09:36Z) - Geometry-Aware Supertagging with Heterogeneous Dynamic Convolutions [0.7868449549351486]
We revisit constructive supertagging from a graph-theoretic perspective.
We propose a framework based on heterogeneous dynamic graph convolutions.
We test our approach on a number of categorial grammar datasets spanning different languages.
arXiv Detail & Related papers (2022-03-23T07:07:11Z) - Rule Augmented Unsupervised Constituency Parsing [11.775897250472116]
We propose an approach that utilizes very generic linguistic knowledge of the language present in the form of syntactic rules.
We achieve new state-of-the-art results on two benchmarks datasets, MNLI and WSJ.
arXiv Detail & Related papers (2021-05-21T08:06:11Z) - VLGrammar: Grounded Grammar Induction of Vision and Language [86.88273769411428]
We study grounded grammar induction of vision and language in a joint learning framework.
We present VLGrammar, a method that uses compound probabilistic context-free grammars (compound PCFGs) to induce the language grammar and the image grammar simultaneously.
arXiv Detail & Related papers (2021-03-24T04:05:08Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - The Return of Lexical Dependencies: Neural Lexicalized PCFGs [103.41187595153652]
We present novel neural models of lexicalized PCFGs which allow us to overcome sparsity problems.
Experiments demonstrate that this unified framework results in stronger results on both representations than achieved when either formalism alone.
arXiv Detail & Related papers (2020-07-29T22:12:49Z) - On embedding Lambek calculus into commutative categorial grammars [0.0]
We consider tensor grammars, which are an example of commutative" grammars, based on the classical (rather than intuitionistic) linear logic.
The basic ingredient are tensor terms, which can be seen as encoding and generalizing proof-nets.
arXiv Detail & Related papers (2020-05-20T14:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.