XLTime: A Cross-Lingual Knowledge Transfer Framework for Temporal
Expression Extraction
- URL: http://arxiv.org/abs/2205.01757v1
- Date: Tue, 3 May 2022 20:00:42 GMT
- Title: XLTime: A Cross-Lingual Knowledge Transfer Framework for Temporal
Expression Extraction
- Authors: Yuwei Cao, William Groves, Tanay Kumar Saha, Joel R. Tetreault, Alex
Jaimes, Hao Peng, and Philip S. Yu
- Abstract summary: Temporal Expression Extraction (TEE) is essential for understanding time in natural language.
To date, work in this area has mostly focused on English as there is a scarcity of labeled data for other languages.
We propose XLTime, a novel framework for multilingual TEE.
- Score: 63.39190486298887
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Temporal Expression Extraction (TEE) is essential for understanding time in
natural language. It has applications in Natural Language Processing (NLP)
tasks such as question answering, information retrieval, and causal inference.
To date, work in this area has mostly focused on English as there is a scarcity
of labeled data for other languages. We propose XLTime, a novel framework for
multilingual TEE. XLTime works on top of pre-trained language models and
leverages multi-task learning to prompt cross-language knowledge transfer both
from English and within the non-English languages. XLTime alleviates problems
caused by a shortage of data in the target language. We apply XLTime with
different language models and show that it outperforms the previous automatic
SOTA methods on French, Spanish, Portuguese, and Basque, by large margins.
XLTime also closes the gap considerably on the handcrafted HeidelTime method.
Related papers
- Crosslingual Capabilities and Knowledge Barriers in Multilingual Large Language Models [62.91524967852552]
Large language models (LLMs) are typically multilingual due to pretraining on diverse multilingual corpora.
But can these models relate corresponding concepts across languages, effectively being crosslingual?
This study evaluates six state-of-the-art LLMs on inherently crosslingual tasks.
arXiv Detail & Related papers (2024-06-23T15:15:17Z) - Cross-Lingual Transfer for Natural Language Inference via Multilingual Prompt Translator [104.63314132355221]
Cross-lingual transfer with prompt learning has shown promising effectiveness.
We propose a novel framework, Multilingual Prompt Translator (MPT)
MPT is more prominent compared with vanilla prompting when transferring to languages quite distinct from source language.
arXiv Detail & Related papers (2024-03-19T03:35:18Z) - LEIA: Facilitating Cross-lingual Knowledge Transfer in Language Models with Entity-based Data Augmentation [21.980770995466134]
We introduce LEIA, a language adaptation tuning method that utilizes Wikipedia entity names aligned across languages.
This method involves augmenting the target language corpus with English entity names and training the model using left-to-right language modeling.
arXiv Detail & Related papers (2024-02-18T07:24:34Z) - Soft Language Clustering for Multilingual Model Pre-training [57.18058739931463]
We propose XLM-P, which contextually retrieves prompts as flexible guidance for encoding instances conditionally.
Our XLM-P enables (1) lightweight modeling of language-invariant and language-specific knowledge across languages, and (2) easy integration with other multilingual pre-training methods.
arXiv Detail & Related papers (2023-06-13T08:08:08Z) - ERNIE-Code: Beyond English-Centric Cross-lingual Pretraining for
Programming Languages [37.60016772021422]
Software engineers working with the same programming language (PL) may speak different natural languages (NLs) and vice versa.
Recent studies have demonstrated the effectiveness of generative pre-training in computer programs, yet they are always English-centric.
We release ERNIE-Code, a unified pre-trained language model for 116 NLs and 6 PLs.
arXiv Detail & Related papers (2022-12-13T17:21:44Z) - One Question Answering Model for Many Languages with Cross-lingual Dense
Passage Retrieval [39.061900747689094]
CORA is a Cross-lingual Open-Retrieval Answer Generation model.
It can answer questions across many languages even when language-specific annotated data or knowledge sources are unavailable.
arXiv Detail & Related papers (2021-07-26T06:02:54Z) - VECO: Variable and Flexible Cross-lingual Pre-training for Language
Understanding and Generation [77.82373082024934]
We plug a cross-attention module into the Transformer encoder to explicitly build the interdependence between languages.
It can effectively avoid the degeneration of predicting masked words only conditioned on the context in its own language.
The proposed cross-lingual model delivers new state-of-the-art results on various cross-lingual understanding tasks of the XTREME benchmark.
arXiv Detail & Related papers (2020-10-30T03:41:38Z) - FILTER: An Enhanced Fusion Method for Cross-lingual Language
Understanding [85.29270319872597]
We propose an enhanced fusion method that takes cross-lingual data as input for XLM finetuning.
During inference, the model makes predictions based on the text input in the target language and its translation in the source language.
To tackle this issue, we propose an additional KL-divergence self-teaching loss for model training, based on auto-generated soft pseudo-labels for translated text in the target language.
arXiv Detail & Related papers (2020-09-10T22:42:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.