Enhanced Universal Dependency Parsing with Second-Order Inference and
Mixture of Training Data
- URL: http://arxiv.org/abs/2006.01414v3
- Date: Wed, 2 Jun 2021 03:07:41 GMT
- Title: Enhanced Universal Dependency Parsing with Second-Order Inference and
Mixture of Training Data
- Authors: Xinyu Wang, Yong Jiang, Kewei Tu
- Abstract summary: This paper presents the system used in our submission to the textitIWPT 2020 Shared Task.
For the low-resource Tamil corpus, we specially mixed the training data of Tamil with other languages and significantly improved the performance of Tamil.
- Score: 48.8386313914471
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: This paper presents the system used in our submission to the \textit{IWPT
2020 Shared Task}. Our system is a graph-based parser with second-order
inference. For the low-resource Tamil corpus, we specially mixed the training
data of Tamil with other languages and significantly improved the performance
of Tamil. Due to our misunderstanding of the submission requirements, we
submitted graphs that are not connected, which makes our system only rank
\textbf{6th} over 10 teams. However, after we fixed this problem, our system is
0.6 ELAS higher than the team that ranked \textbf{1st} in the official results.
Related papers
- BJTU-WeChat's Systems for the WMT22 Chat Translation Task [66.81525961469494]
This paper introduces the joint submission of the Beijing Jiaotong University and WeChat AI to the WMT'22 chat translation task for English-German.
Based on the Transformer, we apply several effective variants.
Our systems achieve 0.810 and 0.946 COMET scores.
arXiv Detail & Related papers (2022-11-28T02:35:04Z) - Alibaba-Translate China's Submission for WMT 2022 Quality Estimation
Shared Task [80.22825549235556]
We present our submission to the sentence-level MQM benchmark at Quality Estimation Shared Task, named UniTE.
Specifically, our systems employ the framework of UniTE, which combined three types of input formats during training with a pre-trained language model.
Results show that our models reach 1st overall ranking in the Multilingual and English-Russian settings, and 2nd overall ranking in English-German and Chinese-English settings.
arXiv Detail & Related papers (2022-10-18T08:55:27Z) - Tencent AI Lab - Shanghai Jiao Tong University Low-Resource Translation
System for the WMT22 Translation Task [49.916963624249355]
This paper describes Tencent AI Lab - Shanghai Jiao Tong University (TAL-SJTU) Low-Resource Translation systems for the WMT22 shared task.
We participate in the general translation task on English$Leftrightarrow$Livonian.
Our system is based on M2M100 with novel techniques that adapt it to the target language pair.
arXiv Detail & Related papers (2022-10-17T04:34:09Z) - PSG@HASOC-Dravidian CodeMixFIRE2021: Pretrained Transformers for
Offensive Language Identification in Tanglish [0.0]
This paper describes the system submitted to Dravidian-Codemix-HASOC2021: Hate Speech and Offensive Language Identification in Dravidian languages.
This task aims to identify offensive content in code-mixed comments/posts in Dravidian languages collected from social media.
arXiv Detail & Related papers (2021-10-06T15:23:40Z) - Netmarble AI Center's WMT21 Automatic Post-Editing Shared Task
Submission [6.043109546012043]
This paper describes Netmarble's submission to WMT21 Automatic Post-Editing (APE) Shared Task for the English-German language pair.
Facebook Fair's WMT19 news translation model was chosen to engage the large and powerful pre-trained neural networks.
For better performance, we leverage external translations as augmented machine translation (MT) during the post-training and fine-tuning.
arXiv Detail & Related papers (2021-09-14T08:21:18Z) - The USYD-JD Speech Translation System for IWSLT 2021 [85.64797317290349]
This paper describes the University of Sydney& JD's joint submission of the IWSLT 2021 low resource speech translation task.
We trained our models with the officially provided ASR and MT datasets.
To achieve better translation performance, we explored the most recent effective strategies, including back translation, knowledge distillation, multi-feature reranking and transductive finetuning.
arXiv Detail & Related papers (2021-07-24T09:53:34Z) - Facebook AI's WMT20 News Translation Task Submission [69.92594751788403]
This paper describes Facebook AI's submission to WMT20 shared news translation task.
We focus on the low resource setting and participate in two language pairs, Tamil -> English and Inuktitut -> English.
We approach the low resource problem using two main strategies, leveraging all available data and adapting the system to the target news domain.
arXiv Detail & Related papers (2020-11-16T21:49:00Z) - K{\o}psala: Transition-Based Graph Parsing via Efficient Training and
Effective Encoding [13.490365811869719]
We present Kopsala, the Copenhagen-Uppsala system for the Enhanced Universal Dependencies Shared Task at IWPT 2020.
Our system is a pipeline consisting of off-the-shelf models for everything but enhanced parsing, and for the latter, a transition-based graphencies adapted from Che et al.
Our demonstrates that a unified pipeline is effective for both Representation Parsing and Enhanced Universal Dependencies, according to average ELAS.
arXiv Detail & Related papers (2020-05-25T13:17:09Z) - Semi-Supervised Neural System for Tagging, Parsing and Lematization [1.6752182911522522]
This paper describes the ICS PAS system which took part in CoNLL 2018 shared task on Multilingual Parsing from Raw Text to Universal Dependencies.
The system consists of jointly trained tagger, lemmatizer, and dependency which are based on features extracted by a biLSTM network.
arXiv Detail & Related papers (2020-04-26T18:29:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.