On the Relation between Syntactic Divergence and Zero-Shot Performance
- URL: http://arxiv.org/abs/2110.04644v1
- Date: Sat, 9 Oct 2021 21:09:21 GMT
- Title: On the Relation between Syntactic Divergence and Zero-Shot Performance
- Authors: Ofir Arviv, Dmitry Nikolaev, Taelin Karidi and Omri Abend
- Abstract summary: We take the transfer of Universal Dependencies (UD) parsing from English to a diverse set of languages and conduct two sets of experiments.
We analyze zero-shot performance based on the extent to which English source edges are preserved in translation.
In both sets of experiments, our results suggest a strong relation between cross-lingual stability and zero-shot parsing performance.
- Score: 22.195133438732633
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We explore the link between the extent to which syntactic relations are
preserved in translation and the ease of correctly constructing a parse tree in
a zero-shot setting. While previous work suggests such a relation, it tends to
focus on the macro level and not on the level of individual edges-a gap we aim
to address. As a test case, we take the transfer of Universal Dependencies (UD)
parsing from English to a diverse set of languages and conduct two sets of
experiments. In one, we analyze zero-shot performance based on the extent to
which English source edges are preserved in translation. In another, we apply
three linguistically motivated transformations to UD, creating more
cross-lingually stable versions of it, and assess their zero-shot parsability.
In order to compare parsing performance across different schemes, we perform
extrinsic evaluation on the downstream task of cross-lingual relation
extraction (RE) using a subset of a popular English RE benchmark translated to
Russian and Korean. In both sets of experiments, our results suggest a strong
relation between cross-lingual stability and zero-shot parsing performance.
Related papers
- Syntactic Language Change in English and German: Metrics, Parsers, and Convergences [56.47832275431858]
The current paper looks at diachronic trends in syntactic language change in both English and German, using corpora of parliamentary debates from the last c. 160 years.
We base our observations on five dependencys, including the widely used Stanford Core as well as 4 newer alternatives.
We show that changes in syntactic measures seem to be more frequent at the tails of sentence length distributions.
arXiv Detail & Related papers (2024-02-18T11:46:16Z) - Narrowing the Gap between Zero- and Few-shot Machine Translation by
Matching Styles [53.92189950211852]
Large language models have demonstrated their ability to generalize to machine translation using zero- and few-shot examples with in-context learning.
In this paper, we investigate the factors contributing to this gap and find that this gap can largely be closed (for about 70%) by matching the writing styles of the target corpus.
arXiv Detail & Related papers (2023-11-04T03:18:45Z) - Optimal Transport Posterior Alignment for Cross-lingual Semantic Parsing [68.47787275021567]
Cross-lingual semantic parsing transfers parsing capability from a high-resource language (e.g., English) to low-resource languages with scarce training data.
We propose a new approach to cross-lingual semantic parsing by explicitly minimizing cross-lingual divergence between latent variables using Optimal Transport.
arXiv Detail & Related papers (2023-07-09T04:52:31Z) - Improving Zero-shot Multilingual Neural Machine Translation by
Leveraging Cross-lingual Consistency Regularization [46.09132547431629]
The multilingual neural machine translation (NMT) model has a promising capability of zero-shot translation.
This paper introduces a cross-lingual consistency regularization, CrossConST, to bridge the representation gap among different languages.
arXiv Detail & Related papers (2023-05-12T08:32:18Z) - CROP: Zero-shot Cross-lingual Named Entity Recognition with Multilingual
Labeled Sequence Translation [113.99145386490639]
Cross-lingual NER can transfer knowledge between languages via aligned cross-lingual representations or machine translation results.
We propose a Cross-lingual Entity Projection framework (CROP) to enable zero-shot cross-lingual NER.
We adopt a multilingual labeled sequence translation model to project the tagged sequence back to the target language and label the target raw sentence.
arXiv Detail & Related papers (2022-10-13T13:32:36Z) - Aligned Weight Regularizers for Pruning Pretrained Neural Networks [6.000551438232907]
We show that there is a clear performance discrepancy in magnitude-based pruning when comparing standard supervised learning to the zero-shot setting.
We propose two weight regularizers that aim to maximize the alignment between units of pruned and unpruned networks.
arXiv Detail & Related papers (2022-04-04T11:06:42Z) - On Cross-Lingual Retrieval with Multilingual Text Encoders [51.60862829942932]
We study the suitability of state-of-the-art multilingual encoders for cross-lingual document and sentence retrieval tasks.
We benchmark their performance in unsupervised ad-hoc sentence- and document-level CLIR experiments.
We evaluate multilingual encoders fine-tuned in a supervised fashion (i.e., we learn to rank) on English relevance data in a series of zero-shot language and domain transfer CLIR experiments.
arXiv Detail & Related papers (2021-12-21T08:10:27Z) - Subword Segmentation and a Single Bridge Language Affect Zero-Shot
Neural Machine Translation [36.4055239280145]
We investigate zero-shot performance of a multilingual EN$leftrightarrow$FR,CS,DE,FI system trained on WMT data.
We observe a bias towards copying the source in zero-shot translation, and investigate how the choice of subword segmentation affects this bias.
arXiv Detail & Related papers (2020-11-03T13:45:54Z) - GATE: Graph Attention Transformer Encoder for Cross-lingual Relation and
Event Extraction [107.8262586956778]
We introduce graph convolutional networks (GCNs) with universal dependency parses to learn language-agnostic sentence representations.
GCNs struggle to model words with long-range dependencies or are not directly connected in the dependency tree.
We propose to utilize the self-attention mechanism to learn the dependencies between words with different syntactic distances.
arXiv Detail & Related papers (2020-10-06T20:30:35Z) - Fine-Grained Analysis of Cross-Linguistic Syntactic Divergences [18.19093600136057]
We propose a framework for extracting divergence patterns for any language pair from a parallel corpus.
We show that our framework provides a detailed picture of cross-language divergences, generalizes previous approaches, and lends itself to full automation.
arXiv Detail & Related papers (2020-05-07T13:05:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.