Exposing Cross-Lingual Lexical Knowledge from Multilingual Sentence
Encoders
- URL: http://arxiv.org/abs/2205.00267v1
- Date: Sat, 30 Apr 2022 13:23:16 GMT
- Title: Exposing Cross-Lingual Lexical Knowledge from Multilingual Sentence
Encoders
- Authors: Ivan Vuli\'c, Goran Glava\v{s}, Fangyu Liu, Nigel Collier, Edoardo
Maria Ponti, Anna Korhonen
- Abstract summary: We probe multilingual language models for the amount of cross-lingual lexical knowledge stored in their parameters, and compare them against the original multilingual LMs.
We also devise a novel method to expose this knowledge by additionally fine-tuning multilingual models.
We report substantial gains on standard benchmarks.
- Score: 85.80950708769923
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Pretrained multilingual language models (LMs) can be successfully transformed
into multilingual sentence encoders (SEs; e.g., LaBSE, xMPNET) via additional
fine-tuning or model distillation on parallel data. However, it remains
uncertain how to best leverage their knowledge to represent sub-sentence
lexical items (i.e., words and phrases) in cross-lingual lexical tasks. In this
work, we probe these SEs for the amount of cross-lingual lexical knowledge
stored in their parameters, and compare them against the original multilingual
LMs. We also devise a novel method to expose this knowledge by additionally
fine-tuning multilingual models through inexpensive contrastive learning
procedure, requiring only a small amount of word translation pairs. We evaluate
our method on bilingual lexical induction (BLI), cross-lingual lexical semantic
similarity, and cross-lingual entity linking, and report substantial gains on
standard benchmarks (e.g., +10 Precision@1 points in BLI), validating that the
SEs such as LaBSE can be 'rewired' into effective cross-lingual lexical
encoders. Moreover, we show that resulting representations can be successfully
interpolated with static embeddings from cross-lingual word embedding spaces to
further boost the performance in lexical tasks. In sum, our approach provides
an effective tool for exposing and harnessing multilingual lexical knowledge
'hidden' in multilingual sentence encoders.
Related papers
- Crosslingual Capabilities and Knowledge Barriers in Multilingual Large Language Models [62.91524967852552]
Large language models (LLMs) are typically multilingual due to pretraining on diverse multilingual corpora.
But can these models relate corresponding concepts across languages, effectively being crosslingual?
This study evaluates six state-of-the-art LLMs on inherently crosslingual tasks.
arXiv Detail & Related papers (2024-06-23T15:15:17Z) - Improving Multi-lingual Alignment Through Soft Contrastive Learning [9.454626745893798]
We propose a novel method to align multi-lingual embeddings based on the similarity of sentences measured by a pre-trained mono-lingual embedding model.
Given translation sentence pairs, we train a multi-lingual model in a way that the similarity between cross-lingual embeddings follows the similarity of sentences measured at the mono-lingual teacher model.
arXiv Detail & Related papers (2024-05-25T09:46:07Z) - Translate to Disambiguate: Zero-shot Multilingual Word Sense
Disambiguation with Pretrained Language Models [67.19567060894563]
Pretrained Language Models (PLMs) learn rich cross-lingual knowledge and can be finetuned to perform well on diverse tasks.
We present a new study investigating how well PLMs capture cross-lingual word sense with Contextual Word-Level Translation (C-WLT)
We find that as the model size increases, PLMs encode more cross-lingual word sense knowledge and better use context to improve WLT performance.
arXiv Detail & Related papers (2023-04-26T19:55:52Z) - Efficiently Aligned Cross-Lingual Transfer Learning for Conversational
Tasks using Prompt-Tuning [98.60739735409243]
Cross-lingual transfer of language models trained on high-resource languages like English has been widely studied for many NLP tasks.
We introduce XSGD for cross-lingual alignment pretraining, a parallel and large-scale multilingual conversation dataset.
To facilitate aligned cross-lingual representations, we develop an efficient prompt-tuning-based method for learning alignment prompts.
arXiv Detail & Related papers (2023-04-03T18:46:01Z) - Multilingual Representation Distillation with Contrastive Learning [20.715534360712425]
We integrate contrastive learning into multilingual representation distillation and use it for quality estimation of parallel sentences.
We validate our approach with multilingual similarity search and corpus filtering tasks.
arXiv Detail & Related papers (2022-10-10T22:27:04Z) - Massively Multilingual Lexical Specialization of Multilingual
Transformers [18.766379322798837]
We show that massively multilingual lexical specialization brings substantial gains in two standard cross-lingual lexical tasks.
We observe gains for languages unseen in specialization, indicating that multilingual lexical specialization enables generalization to languages with no lexical constraints.
arXiv Detail & Related papers (2022-08-01T17:47:03Z) - Learning Contextualised Cross-lingual Word Embeddings and Alignments for
Extremely Low-Resource Languages Using Parallel Corpora [63.5286019659504]
We propose a new approach for learning contextualised cross-lingual word embeddings based on a small parallel corpus.
Our method obtains word embeddings via an LSTM encoder-decoder model that simultaneously translates and reconstructs an input sentence.
arXiv Detail & Related papers (2020-10-27T22:24:01Z) - FILTER: An Enhanced Fusion Method for Cross-lingual Language
Understanding [85.29270319872597]
We propose an enhanced fusion method that takes cross-lingual data as input for XLM finetuning.
During inference, the model makes predictions based on the text input in the target language and its translation in the source language.
To tackle this issue, we propose an additional KL-divergence self-teaching loss for model training, based on auto-generated soft pseudo-labels for translated text in the target language.
arXiv Detail & Related papers (2020-09-10T22:42:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.