Cross-Lingual Transfer for Natural Language Inference via Multilingual Prompt Translator
- URL: http://arxiv.org/abs/2403.12407v1
- Date: Tue, 19 Mar 2024 03:35:18 GMT
- Title: Cross-Lingual Transfer for Natural Language Inference via Multilingual Prompt Translator
- Authors: Xiaoyu Qiu, Yuechen Wang, Jiaxin Shi, Wengang Zhou, Houqiang Li,
- Abstract summary: Cross-lingual transfer with prompt learning has shown promising effectiveness.
We propose a novel framework, Multilingual Prompt Translator (MPT)
MPT is more prominent compared with vanilla prompting when transferring to languages quite distinct from source language.
- Score: 104.63314132355221
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Based on multilingual pre-trained models, cross-lingual transfer with prompt learning has shown promising effectiveness, where soft prompt learned in a source language is transferred to target languages for downstream tasks, particularly in the low-resource scenario. To efficiently transfer soft prompt, we propose a novel framework, Multilingual Prompt Translator (MPT), where a multilingual prompt translator is introduced to properly process crucial knowledge embedded in prompt by changing language knowledge while retaining task knowledge. Concretely, we first train prompt in source language and employ translator to translate it into target prompt. Besides, we extend an external corpus as auxiliary data, on which an alignment task for predicted answer probability is designed to convert language knowledge, thereby equipping target prompt with multilingual knowledge. In few-shot settings on XNLI, MPT demonstrates superiority over baselines by remarkable improvements. MPT is more prominent compared with vanilla prompting when transferring to languages quite distinct from source language.
Related papers
- Soft Language Prompts for Language Transfer [0.0]
Cross-lingual knowledge transfer, especially between high- and low-resource languages, remains challenging in natural language processing (NLP)
This study offers insights for improving cross-lingual NLP applications through the combination of parameter-efficient fine-tuning methods.
arXiv Detail & Related papers (2024-07-02T14:50:03Z) - Self-Augmentation Improves Zero-Shot Cross-Lingual Transfer [92.80671770992572]
Cross-lingual transfer is a central task in multilingual NLP.
Earlier efforts on this task use parallel corpora, bilingual dictionaries, or other annotated alignment data.
We propose a simple yet effective method, SALT, to improve the zero-shot cross-lingual transfer.
arXiv Detail & Related papers (2023-09-19T19:30:56Z) - Soft Language Clustering for Multilingual Model Pre-training [57.18058739931463]
We propose XLM-P, which contextually retrieves prompts as flexible guidance for encoding instances conditionally.
Our XLM-P enables (1) lightweight modeling of language-invariant and language-specific knowledge across languages, and (2) easy integration with other multilingual pre-training methods.
arXiv Detail & Related papers (2023-06-13T08:08:08Z) - Enhancing Cross-lingual Natural Language Inference by Soft Prompting
with Multilingual Verbalizer [52.46740830977898]
Cross-lingual natural language inference is a fundamental problem in cross-lingual language understanding.
We propose a novel Soft prompt learning framework with the Multilingual Verbalizer (SoftMV) for XNLI.
arXiv Detail & Related papers (2023-05-22T06:31:29Z) - Zero-shot Cross-lingual Transfer of Prompt-based Tuning with a Unified
Multilingual Prompt [98.26682501616024]
We propose a novel model that uses a unified prompt for all languages, called UniPrompt.
The unified prompt is computation by a multilingual PLM to produce language-independent representation.
Our proposed methods can significantly outperform the strong baselines across different languages.
arXiv Detail & Related papers (2022-02-23T11:57:52Z) - Syntax-augmented Multilingual BERT for Cross-lingual Transfer [37.99210035238424]
This work shows that explicitly providing language syntax and training mBERT helps cross-lingual transfer.
Experiment results show that syntax-augmented mBERT improves cross-lingual transfer on popular benchmarks.
arXiv Detail & Related papers (2021-06-03T21:12:50Z) - FILTER: An Enhanced Fusion Method for Cross-lingual Language
Understanding [85.29270319872597]
We propose an enhanced fusion method that takes cross-lingual data as input for XLM finetuning.
During inference, the model makes predictions based on the text input in the target language and its translation in the source language.
To tackle this issue, we propose an additional KL-divergence self-teaching loss for model training, based on auto-generated soft pseudo-labels for translated text in the target language.
arXiv Detail & Related papers (2020-09-10T22:42:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.