K{\o}psala: Transition-Based Graph Parsing via Efficient Training and
Effective Encoding
- URL: http://arxiv.org/abs/2005.12094v2
- Date: Tue, 2 Jun 2020 11:45:56 GMT
- Title: K{\o}psala: Transition-Based Graph Parsing via Efficient Training and
Effective Encoding
- Authors: Daniel Hershcovich, Miryam de Lhoneux, Artur Kulmizev, Elham Pejhan,
Joakim Nivre
- Abstract summary: We present Kopsala, the Copenhagen-Uppsala system for the Enhanced Universal Dependencies Shared Task at IWPT 2020.
Our system is a pipeline consisting of off-the-shelf models for everything but enhanced parsing, and for the latter, a transition-based graphencies adapted from Che et al.
Our demonstrates that a unified pipeline is effective for both Representation Parsing and Enhanced Universal Dependencies, according to average ELAS.
- Score: 13.490365811869719
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present K{\o}psala, the Copenhagen-Uppsala system for the Enhanced
Universal Dependencies Shared Task at IWPT 2020. Our system is a pipeline
consisting of off-the-shelf models for everything but enhanced graph parsing,
and for the latter, a transition-based graph parser adapted from Che et al.
(2019). We train a single enhanced parser model per language, using gold
sentence splitting and tokenization for training, and rely only on tokenized
surface forms and multilingual BERT for encoding. While a bug introduced just
before submission resulted in a severe drop in precision, its post-submission
fix would bring us to 4th place in the official ranking, according to average
ELAS. Our parser demonstrates that a unified pipeline is effective for both
Meaning Representation Parsing and Enhanced Universal Dependencies.
Related papers
- Breaking the Script Barrier in Multilingual Pre-Trained Language Models with Transliteration-Based Post-Training Alignment [50.27950279695363]
The transfer performance is often hindered when a low-resource target language is written in a different script than the high-resource source language.
Inspired by recent work that uses transliteration to address this problem, our paper proposes a transliteration-based post-pretraining alignment (PPA) method.
arXiv Detail & Related papers (2024-06-28T08:59:24Z) - A Data-Driven Representation for Sign Language Production [26.520016084139964]
Sign Language Production aims to automatically translate spoken language sentences into continuous sequences of sign language.
Current state-of-the-art approaches rely on scarce linguistic resources to work.
This paper introduces an innovative solution by transforming the continuous pose generation problem into a discrete sequence generation problem.
arXiv Detail & Related papers (2024-04-17T15:52:38Z) - Dual-Alignment Pre-training for Cross-lingual Sentence Embedding [79.98111074307657]
We propose a dual-alignment pre-training (DAP) framework for cross-lingual sentence embedding.
We introduce a novel representation translation learning (RTL) task, where the model learns to use one-side contextualized token representation to reconstruct its translation counterpart.
Our approach can significantly improve sentence embedding.
arXiv Detail & Related papers (2023-05-16T03:53:30Z) - A Simple Multi-Modality Transfer Learning Baseline for Sign Language
Translation [54.29679610921429]
Existing sign language datasets contain only about 10K-20K pairs of sign videos, gloss annotations and texts.
Data is thus a bottleneck for training effective sign language translation models.
This simple baseline surpasses the previous state-of-the-art results on two sign language translation benchmarks.
arXiv Detail & Related papers (2022-03-08T18:59:56Z) - Improving Multilingual Translation by Representation and Gradient
Regularization [82.42760103045083]
We propose a joint approach to regularize NMT models at both representation-level and gradient-level.
Our results demonstrate that our approach is highly effective in both reducing off-target translation occurrences and improving zero-shot translation performance.
arXiv Detail & Related papers (2021-09-10T10:52:21Z) - Translate & Fill: Improving Zero-Shot Multilingual Semantic Parsing with
Synthetic Data [2.225882303328135]
We propose a novel Translate-and-Fill (TaF) method to produce silver training data for a multilingual semantic parsing task.
Experimental results on three multilingual semantic parsing datasets show that data augmentation with TaF reaches accuracies competitive with similar systems.
arXiv Detail & Related papers (2021-09-09T14:51:11Z) - TGIF: Tree-Graph Integrated-Format Parser for Enhanced UD with Two-Stage
Generic- to Individual-Language Finetuning [18.71574180551552]
We present our contribution to the IWPT 2021 shared task on parsing into enhanced Universal Dependencies.
Our main system component is a hybrid tree-graph that integrates predictions of spanning trees for the enhanced graphs with additional graph edges not present in the spanning trees.
arXiv Detail & Related papers (2021-07-14T18:00:08Z) - The DCU-EPFL Enhanced Dependency Parser at the IWPT 2021 Shared Task [19.98425994656106]
We describe the multitask-EPFL submission to the IWPT 2021 Shared Task on Parsing into Enhanced Universal Dependencies.
The task involves parsing Enhanced graphs, which are an extension of the basic dependency trees designed to be more facilitative towards representing semantic structure.
evaluation is carried out on 29 treebanks in 17 languages and participants are required to parse the data from each language starting from raw strings.
arXiv Detail & Related papers (2021-07-05T12:42:59Z) - The ADAPT Enhanced Dependency Parser at the IWPT 2020 Shared Task [12.226699055857182]
We describe the ADAPT system for the 2020 IWPT Shared Task on parsing enhanced Universal Dependencies in 17 languages.
We implement a pipeline approach using UDPipe and UDPipe-future to provide initial levels of annotation.
For the majority of languages, a semantic dependency can be successfully applied to the task of parsing enhanced dependencies.
arXiv Detail & Related papers (2020-09-03T14:43:04Z) - Towards Instance-Level Parser Selection for Cross-Lingual Transfer of
Dependency Parsers [59.345145623931636]
We argue for a novel cross-lingual transfer paradigm: instance-level selection (ILPS)
We present a proof-of-concept study focused on instance-level selection in the framework of delexicalized transfer.
arXiv Detail & Related papers (2020-04-16T13:18:55Z) - Sign Language Transformers: Joint End-to-end Sign Language Recognition
and Translation [59.38247587308604]
We introduce a novel transformer based architecture that jointly learns Continuous Sign Language Recognition and Translation.
We evaluate the recognition and translation performances of our approaches on the challenging RWTH-PHOENIX-Weather-2014T dataset.
Our translation networks outperform both sign video to spoken language and gloss to spoken language translation models.
arXiv Detail & Related papers (2020-03-30T21:35:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.