Discontinuous Grammar as a Foreign Language
- URL: http://arxiv.org/abs/2110.10431v1
- Date: Wed, 20 Oct 2021 08:58:02 GMT
- Title: Discontinuous Grammar as a Foreign Language
- Authors: Daniel Fern\'andez-Gonz\'alez and Carlos G\'omez-Rodr\'iguez
- Abstract summary: We extend the framework of sequence-to-sequence models for constituent parsing.
We design several novelizations that can fully produce discontinuities.
For the first time, we test a sequence-to-sequence model on the main discontinuous benchmarks.
- Score: 0.7412445894287709
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In order to achieve deep natural language understanding, syntactic
constituent parsing is a vital step, highly demanded by many artificial
intelligence systems to process both text and speech. One of the most recent
proposals is the use of standard sequence-to-sequence models to perform
constituent parsing as a machine translation task, instead of applying
task-specific parsers. While they show a competitive performance, these
text-to-parse transducers are still lagging behind classic techniques in terms
of accuracy, coverage and speed. To close the gap, we here extend the framework
of sequence-to-sequence models for constituent parsing, not only by providing a
more powerful neural architecture for improving their performance, but also by
enlarging their coverage to handle the most complex syntactic phenomena:
discontinuous structures. To that end, we design several novel linearizations
that can fully produce discontinuities and, for the first time, we test a
sequence-to-sequence model on the main discontinuous benchmarks, obtaining
competitive results on par with task-specific discontinuous constituent parsers
and achieving state-of-the-art scores on the (discontinuous) English Penn
Treebank.
Related papers
- Activity Grammars for Temporal Action Segmentation [71.03141719666972]
temporal action segmentation aims at translating an untrimmed activity video into a sequence of action segments.
This paper introduces an effective activity grammar to guide neural predictions for temporal action segmentation.
Experimental results demonstrate that our method significantly improves temporal action segmentation in terms of both performance and interpretability.
arXiv Detail & Related papers (2023-12-07T12:45:33Z) - Structural generalization is hard for sequence-to-sequence models [85.0087839979613]
Sequence-to-sequence (seq2seq) models have been successful across many NLP tasks.
Recent work on compositional generalization has shown that seq2seq models achieve very low accuracy in generalizing to linguistic structures that were not seen in training.
arXiv Detail & Related papers (2022-10-24T09:03:03Z) - Sentence Representation Learning with Generative Objective rather than
Contrastive Objective [86.01683892956144]
We propose a novel generative self-supervised learning objective based on phrase reconstruction.
Our generative learning achieves powerful enough performance improvement and outperforms the current state-of-the-art contrastive methods.
arXiv Detail & Related papers (2022-10-16T07:47:46Z) - Long Text Generation by Modeling Sentence-Level and Discourse-Level
Coherence [59.51720326054546]
We propose a long text generation model, which can represent the prefix sentences at sentence level and discourse level in the decoding process.
Our model can generate more coherent texts than state-of-the-art baselines.
arXiv Detail & Related papers (2021-05-19T07:29:08Z) - Lexically-constrained Text Generation through Commonsense Knowledge
Extraction and Injection [62.071938098215085]
We focus on the Commongen benchmark, wherein the aim is to generate a plausible sentence for a given set of input concepts.
We propose strategies for enhancing the semantic correctness of the generated text.
arXiv Detail & Related papers (2020-12-19T23:23:40Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Discontinuous Constituent Parsing with Pointer Networks [0.34376560669160383]
discontinuous constituent trees are crucial for representing all grammatical phenomena of languages such as German.
Recent advances in dependency parsing have shown that Pointer Networks excel in efficiently parsing syntactic relations between words in a sentence.
We propose a novel neural network architecture that is able to generate the most accurate discontinuous constituent representations.
arXiv Detail & Related papers (2020-02-05T15:12:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.