Analysis of the Penn Korean Universal Dependency Treebank (PKT-UD):
Manual Revision to Build Robust Parsing Model in Korean
- URL: http://arxiv.org/abs/2005.12898v1
- Date: Tue, 26 May 2020 17:46:46 GMT
- Title: Analysis of the Penn Korean Universal Dependency Treebank (PKT-UD):
Manual Revision to Build Robust Parsing Model in Korean
- Authors: Tae Hwan Oh, Ji Yoon Han, Hyonsu Choe, Seokwon Park, Han He, Jinho D.
Choi, Na-Rae Han, Jena D. Hwang, Hansaem Kim
- Abstract summary: We first open on important issues regarding the Penn Korean Universal Treebank (PKT-UD)
We address these issues by revising the entire corpus manually with the aim of producing cleaner UD annotations.
For compatibility to the rest of UD corpora, we extensively revise the part-of-speech tags and the dependency relations.
- Score: 15.899449418195106
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we first open on important issues regarding the Penn Korean
Universal Treebank (PKT-UD) and address these issues by revising the entire
corpus manually with the aim of producing cleaner UD annotations that are more
faithful to Korean grammar. For compatibility to the rest of UD corpora, we
follow the UDv2 guidelines, and extensively revise the part-of-speech tags and
the dependency relations to reflect morphological features and flexible
word-order aspects in Korean. The original and the revised versions of PKT-UD
are experimented with transformer-based parsing models using biaffine
attention. The parsing model trained on the revised corpus shows a significant
improvement of 3.0% in labeled attachment score over the model trained on the
previous corpus. Our error analysis demonstrates that this revision allows the
parsing model to learn relations more robustly, reducing several critical
errors that used to be made by the previous model.
Related papers
- Linear Recency Bias During Training Improves Transformers' Fit to Reading Times [16.55240473621401]
This paper evaluates a modification of the Transformer model that uses ALiBi, a recency bias added to attention scores.
ALiBi's mixture of slopes -- which determine the rate of memory decay in each attention head -- may play a role in helping models with ALiBi to track different kinds of linguistic dependencies.
arXiv Detail & Related papers (2024-09-17T14:57:51Z) - Thai Universal Dependency Treebank [0.0]
We introduce Thai Universal Dependency Treebank (TUD), a new largest Thai treebank consisting of 3,627 trees annotated in accordance with the Universal Dependencies (UD) framework.
We then benchmark dependency parsing models that incorporate pretrained encoders and train them on Thai-PUD and our TUD.
The results show that most of our models can outperform other models reported in previous papers and provide insight into the optimal choices of components in Thai dependencys.
arXiv Detail & Related papers (2024-05-13T09:48:13Z) - Lost in Translation? Translation Errors and Challenges for Fair Assessment of Text-to-Image Models on Multilingual Concepts [107.32683485639654]
Benchmarks of the multilingual capabilities of text-to-image (T2I) models compare generated images prompted in a test language to an expected image distribution over a concept set.
One such benchmark, "Conceptual Coverage Across Languages" (CoCo-CroLa), assesses the tangible noun inventory of T2I models by prompting them to generate pictures from a concept list translated to seven languages.
We find that this benchmark contains translation errors of varying severity in Spanish, Japanese, and Chinese.
arXiv Detail & Related papers (2024-03-17T05:05:11Z) - Dependency Annotation of Ottoman Turkish with Multilingual BERT [0.0]
This study introduces a pretrained large language model-based annotation methodology for the first dency treebank in Ottoman Turkish.
The resulting treebank will facilitate automated analysis of Ottoman Turkish documents, unlocking the linguistic richness embedded in this historical heritage.
arXiv Detail & Related papers (2024-02-22T17:58:50Z) - Disco-Bench: A Discourse-Aware Evaluation Benchmark for Language
Modelling [70.23876429382969]
We propose a benchmark that can evaluate intra-sentence discourse properties across a diverse set of NLP tasks.
Disco-Bench consists of 9 document-level testsets in the literature domain, which contain rich discourse phenomena.
For linguistic analysis, we also design a diagnostic test suite that can examine whether the target models learn discourse knowledge.
arXiv Detail & Related papers (2023-07-16T15:18:25Z) - Training Language Models with Language Feedback at Scale [50.70091340506957]
We introduce learning from Language Feedback (ILF), a new approach that utilizes more informative language feedback.
ILF consists of three steps that are applied iteratively: first, conditioning the language model on the input, an initial LM output, and feedback to generate refinements.
We show theoretically that ILF can be viewed as Bayesian Inference, similar to Reinforcement Learning from human feedback.
arXiv Detail & Related papers (2023-03-28T17:04:15Z) - Yet Another Format of Universal Dependencies for Korean [4.909210276089872]
morphUD outperforms parsing results for all Korean UD treebanks.
We develop scripts that convert between the original format used by Universal Dependencies and the proposed morpheme-based format automatically.
arXiv Detail & Related papers (2022-09-20T14:21:00Z) - Masked Part-Of-Speech Model: Does Modeling Long Context Help
Unsupervised POS-tagging? [94.68962249604749]
We propose a Masked Part-of-Speech Model (MPoSM) to facilitate flexible dependency modeling.
MPoSM can model arbitrary tag dependency and perform POS induction through the objective of masked POS reconstruction.
We achieve competitive results on both the English Penn WSJ dataset and the universal treebank containing 10 diverse languages.
arXiv Detail & Related papers (2022-06-30T01:43:05Z) - Unsupervised and Few-shot Parsing from Pretrained Language Models [56.33247845224995]
We propose an Unsupervised constituent Parsing model that calculates an Out Association score solely based on the self-attention weight matrix learned in a pretrained language model.
We extend the unsupervised models to few-shot parsing models that use a few annotated trees to learn better linear projection matrices for parsing.
Our few-shot parsing model FPIO trained with only 20 annotated trees outperforms a previous few-shot parsing method trained with 50 annotated trees.
arXiv Detail & Related papers (2022-06-10T10:29:15Z) - KLUE: Korean Language Understanding Evaluation [43.94952771238633]
We introduce Korean Language Understanding Evaluation (KLUE) benchmark.
KLUE is a collection of 8 Korean natural language understanding (NLU) tasks.
We build all of the tasks from scratch from diverse source corpora while respecting copyrights.
arXiv Detail & Related papers (2021-05-20T11:40:30Z) - Rnn-transducer with language bias for end-to-end Mandarin-English
code-switching speech recognition [58.105818353866354]
We propose an improved recurrent neural network transducer (RNN-T) model with language bias to alleviate the problem.
We use the language identities to bias the model to predict the CS points.
This promotes the model to learn the language identity information directly from transcription, and no additional LID model is needed.
arXiv Detail & Related papers (2020-02-19T12:01:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.