A Systematic Comparison of Syntactic Representations of Dependency Parsing
- URL: http://arxiv.org/abs/2503.07142v1
- Date: Mon, 10 Mar 2025 10:13:55 GMT
- Title: A Systematic Comparison of Syntactic Representations of Dependency Parsing
- Authors: Guillaume Wisniewski, Ophélie Lacroix,
- Abstract summary: We pro-pose to convert some specific syntactic constructions observed in the universal dependency treebanks into a so-called more standard representation.<n>We show that the standard'' constructions do not lead systematically to better parsing performance and that the scores vary considerably according to the languages.
- Score: 5.844015313757265
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We compare the performance of a transition-based parser in regards to different annotation schemes. We pro-pose to convert some specific syntactic constructions observed in the universal dependency treebanks into a so-called more standard representation and to evaluate parsing performances over all the languages of the project. We show that the ``standard'' constructions do not lead systematically to better parsing performance and that the scores vary considerably according to the languages.
Related papers
- Predicting Text Preference Via Structured Comparative Reasoning [110.49560164568791]
We introduce SC, a prompting approach that predicts text preferences by generating structured intermediate comparisons.
We select consistent comparisons with a pairwise consistency comparator that ensures each aspect's comparisons clearly distinguish differences between texts.
Our comprehensive evaluations across various NLP tasks, including summarization, retrieval, and automatic rating, demonstrate that SC equips LLMs to achieve state-of-the-art performance in text preference prediction.
arXiv Detail & Related papers (2023-11-14T18:51:38Z) - Assessment of Pre-Trained Models Across Languages and Grammars [7.466159270333272]
We aim to recover constituent and dependency structures by casting parsing as sequence labeling.
Our results show that pre-trained word vectors do not favor constituency representations of syntax over dependencies.
occurrence of a language in the pretraining data is more important than the amount of task data when recovering syntax from the word vectors.
arXiv Detail & Related papers (2023-09-20T09:23:36Z) - Generic Dependency Modeling for Multi-Party Conversation [32.25605889407403]
We present an approach to encoding the dependencies in the form of relative dependency encoding (ReDE)
We show how to implement it in Transformers by modifying the computation of self-attention.
arXiv Detail & Related papers (2023-02-21T13:58:19Z) - BenchCLAMP: A Benchmark for Evaluating Language Models on Syntactic and
Semantic Parsing [55.058258437125524]
We introduce BenchCLAMP, a Benchmark to evaluate Constrained LAnguage Model Parsing.
We benchmark eight language models, including two GPT-3 variants available only through an API.
Our experiments show that encoder-decoder pretrained language models can achieve similar performance or surpass state-of-the-art methods for syntactic and semantic parsing when the model output is constrained to be valid.
arXiv Detail & Related papers (2022-06-21T18:34:11Z) - Multilingual Extraction and Categorization of Lexical Collocations with
Graph-aware Transformers [86.64972552583941]
We put forward a sequence tagging BERT-based model enhanced with a graph-aware transformer architecture, which we evaluate on the task of collocation recognition in context.
Our results suggest that explicitly encoding syntactic dependencies in the model architecture is helpful, and provide insights on differences in collocation typification in English, Spanish and French.
arXiv Detail & Related papers (2022-05-23T16:47:37Z) - On The Ingredients of an Effective Zero-shot Semantic Parser [95.01623036661468]
We analyze zero-shot learning by paraphrasing training examples of canonical utterances and programs from a grammar.
We propose bridging these gaps using improved grammars, stronger paraphrasers, and efficient learning methods.
Our model achieves strong performance on two semantic parsing benchmarks (Scholar, Geo) with zero labeled data.
arXiv Detail & Related papers (2021-10-15T21:41:16Z) - Compositional Generalization in Dependency Parsing [15.953482168182003]
Dependency, however, lacks a compositional parsing benchmark.
We find that increasing compound divergence degrades dependency performance, although not as dramatically as semantic parsing performance.
We identify a number of syntactic structures that drive the dependency's lower performance on the most challenging splits.
arXiv Detail & Related papers (2021-10-13T16:32:24Z) - Infusing Finetuning with Semantic Dependencies [62.37697048781823]
We show that, unlike syntax, semantics is not brought to the surface by today's pretrained models.
We then use convolutional graph encoders to explicitly incorporate semantic parses into task-specific finetuning.
arXiv Detail & Related papers (2020-12-10T01:27:24Z) - Comparison by Conversion: Reverse-Engineering UCCA from Syntax and
Lexical Semantics [29.971739294416714]
Building robust natural language understanding systems will require a clear characterization of whether and how various linguistic meaning representations complement each other.
We evaluate the mapping between meaning representations from different frameworks using two complementary methods: (i) a rule-based converter, and (ii) a supervised delexicalized that parses to one framework using only information from the other as features.
arXiv Detail & Related papers (2020-11-02T09:03:46Z) - Towards Instance-Level Parser Selection for Cross-Lingual Transfer of
Dependency Parsers [59.345145623931636]
We argue for a novel cross-lingual transfer paradigm: instance-level selection (ILPS)
We present a proof-of-concept study focused on instance-level selection in the framework of delexicalized transfer.
arXiv Detail & Related papers (2020-04-16T13:18:55Z) - Transition-Based Dependency Parsing using Perceptron Learner [34.59241394911966]
We tackle transition-based dependency parsing using a Perceptron Learner.
Our proposed model, which adds more relevant features to the Perceptron Learner, outperforms a baseline arc-standard.
arXiv Detail & Related papers (2020-01-22T20:58:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.