Detecting Fine-Grained Cross-Lingual Semantic Divergences without
Supervision by Learning to Rank
- URL: http://arxiv.org/abs/2010.03662v1
- Date: Wed, 7 Oct 2020 21:26:20 GMT
- Title: Detecting Fine-Grained Cross-Lingual Semantic Divergences without
Supervision by Learning to Rank
- Authors: Eleftheria Briakou and Marine Carpuat
- Abstract summary: This work improves the prediction and annotation of fine-grained semantic divergences.
We introduce a training strategy for multilingual BERT models by learning to rank synthetic divergent examples of varying granularity.
Learning to rank helps detect fine-grained sentence-level divergences more accurately than a strong sentence-level similarity model.
- Score: 28.910206570036593
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Detecting fine-grained differences in content conveyed in different languages
matters for cross-lingual NLP and multilingual corpora analysis, but it is a
challenging machine learning problem since annotation is expensive and hard to
scale. This work improves the prediction and annotation of fine-grained
semantic divergences. We introduce a training strategy for multilingual BERT
models by learning to rank synthetic divergent examples of varying granularity.
We evaluate our models on the Rationalized English-French Semantic Divergences,
a new dataset released with this work, consisting of English-French
sentence-pairs annotated with semantic divergence classes and token-level
rationales. Learning to rank helps detect fine-grained sentence-level
divergences more accurately than a strong sentence-level similarity model,
while token-level predictions have the potential of further distinguishing
between coarse and fine-grained divergences.
Related papers
- Evaluating and explaining training strategies for zero-shot cross-lingual news sentiment analysis [8.770572911942635]
We introduce novel evaluation datasets in several less-resourced languages.
We experiment with a range of approaches including the use of machine translation.
We show that language similarity is not in itself sufficient for predicting the success of cross-lingual transfer.
arXiv Detail & Related papers (2024-09-30T07:59:41Z) - Improving Multi-lingual Alignment Through Soft Contrastive Learning [9.454626745893798]
We propose a novel method to align multi-lingual embeddings based on the similarity of sentences measured by a pre-trained mono-lingual embedding model.
Given translation sentence pairs, we train a multi-lingual model in a way that the similarity between cross-lingual embeddings follows the similarity of sentences measured at the mono-lingual teacher model.
arXiv Detail & Related papers (2024-05-25T09:46:07Z) - Optimal Transport Posterior Alignment for Cross-lingual Semantic Parsing [68.47787275021567]
Cross-lingual semantic parsing transfers parsing capability from a high-resource language (e.g., English) to low-resource languages with scarce training data.
We propose a new approach to cross-lingual semantic parsing by explicitly minimizing cross-lingual divergence between latent variables using Optimal Transport.
arXiv Detail & Related papers (2023-07-09T04:52:31Z) - VECO 2.0: Cross-lingual Language Model Pre-training with
Multi-granularity Contrastive Learning [56.47303426167584]
We propose a cross-lingual pre-trained model VECO2.0 based on contrastive learning with multi-granularity alignments.
Specifically, the sequence-to-sequence alignment is induced to maximize the similarity of the parallel pairs and minimize the non-parallel pairs.
token-to-token alignment is integrated to bridge the gap between synonymous tokens excavated via the thesaurus dictionary from the other unpaired tokens in a bilingual instance.
arXiv Detail & Related papers (2023-04-17T12:23:41Z) - Beyond Contrastive Learning: A Variational Generative Model for
Multilingual Retrieval [109.62363167257664]
We propose a generative model for learning multilingual text embeddings.
Our model operates on parallel data in $N$ languages.
We evaluate this method on a suite of tasks including semantic similarity, bitext mining, and cross-lingual question retrieval.
arXiv Detail & Related papers (2022-12-21T02:41:40Z) - A Multi-level Supervised Contrastive Learning Framework for Low-Resource
Natural Language Inference [54.678516076366506]
Natural Language Inference (NLI) is a growingly essential task in natural language understanding.
Here we propose a multi-level supervised contrastive learning framework named MultiSCL for low-resource natural language inference.
arXiv Detail & Related papers (2022-05-31T05:54:18Z) - Multi-Level Contrastive Learning for Cross-Lingual Alignment [35.33431650608965]
Cross-language pre-trained models such as multilingual BERT (mBERT) have achieved significant performance in various cross-lingual downstream NLP tasks.
This paper proposes a multi-level contrastive learning framework to further improve the cross-lingual ability of pre-trained models.
arXiv Detail & Related papers (2022-02-26T07:14:20Z) - Modelling Latent Translations for Cross-Lingual Transfer [47.61502999819699]
We propose a new technique that integrates both steps of the traditional pipeline (translation and classification) into a single model.
We evaluate our novel latent translation-based model on a series of multilingual NLU tasks.
We report gains for both zero-shot and few-shot learning setups, up to 2.7 accuracy points on average.
arXiv Detail & Related papers (2021-07-23T17:11:27Z) - Comparative Error Analysis in Neural and Finite-state Models for
Unsupervised Character-level Transduction [34.1177259741046]
We compare the two model classes side by side and find that they tend to make different types of errors even when achieving comparable performance.
We investigate how combining finite-state and sequence-to-sequence models at decoding time affects the output quantitatively and qualitatively.
arXiv Detail & Related papers (2021-06-24T00:09:24Z) - Leveraging Adversarial Training in Self-Learning for Cross-Lingual Text
Classification [52.69730591919885]
We present a semi-supervised adversarial training process that minimizes the maximal loss for label-preserving input perturbations.
We observe significant gains in effectiveness on document and intent classification for a diverse set of languages.
arXiv Detail & Related papers (2020-07-29T19:38:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.