Learning Domain-Specialised Representations for Cross-Lingual Biomedical
Entity Linking
- URL: http://arxiv.org/abs/2105.14398v1
- Date: Sun, 30 May 2021 00:50:00 GMT
- Title: Learning Domain-Specialised Representations for Cross-Lingual Biomedical
Entity Linking
- Authors: Fangyu Liu, Ivan Vuli\'c, Anna Korhonen, Nigel Collier
- Abstract summary: We propose a novel cross-lingual biomedical entity linking task (XL-BEL)
We first investigate the ability of standard knowledge-agnostic as well as knowledge-enhanced monolingual and multilingual LMs beyond the standard monolingual English BEL task.
We then address the challenge of transferring domain-specific knowledge in resource-rich languages to resource-poor ones.
- Score: 66.76141128555099
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Injecting external domain-specific knowledge (e.g., UMLS) into pretrained
language models (LMs) advances their capability to handle specialised in-domain
tasks such as biomedical entity linking (BEL). However, such abundant expert
knowledge is available only for a handful of languages (e.g., English). In this
work, by proposing a novel cross-lingual biomedical entity linking task
(XL-BEL) and establishing a new XL-BEL benchmark spanning 10 typologically
diverse languages, we first investigate the ability of standard
knowledge-agnostic as well as knowledge-enhanced monolingual and multilingual
LMs beyond the standard monolingual English BEL task. The scores indicate large
gaps to English performance. We then address the challenge of transferring
domain-specific knowledge in resource-rich languages to resource-poor ones. To
this end, we propose and evaluate a series of cross-lingual transfer methods
for the XL-BEL task, and demonstrate that general-domain bitext helps propagate
the available English knowledge to languages with little to no in-domain data.
Remarkably, we show that our proposed domain-specific transfer methods yield
consistent gains across all target languages, sometimes up to 20 Precision@1
points, without any in-domain knowledge in the target language, and without any
in-domain parallel data.
Related papers
- Zero-shot Cross-lingual Transfer Learning with Multiple Source and Target Languages for Information Extraction: Language Selection and Adversarial Training [38.19963761398705]
This paper provides a detailed analysis on Cross-Lingual Multi-Transferability (many-to-many transfer learning) for the recent IE corpora.
We first determine the correlation between single-language performance and a wide range of linguistic-based distances.
Next, we investigate the more general zero-shot multi-lingual transfer settings where multiple languages are involved in the training and evaluation processes.
arXiv Detail & Related papers (2024-11-13T17:13:25Z) - Crosslingual Capabilities and Knowledge Barriers in Multilingual Large Language Models [62.91524967852552]
Large language models (LLMs) are typically multilingual due to pretraining on diverse multilingual corpora.
But can these models relate corresponding concepts across languages, effectively being crosslingual?
This study evaluates six state-of-the-art LLMs on inherently crosslingual tasks.
arXiv Detail & Related papers (2024-06-23T15:15:17Z) - LexGen: Domain-aware Multilingual Lexicon Generation [40.97738267067852]
We propose a new model to generate dictionary words for 6 Indian languages in the multi-domain setting.
Our model consists of domain-specific and domain-generic layers that encode information.
We release a new benchmark dataset across 6 Indian languages that span 8 diverse domains.
arXiv Detail & Related papers (2024-05-18T07:02:43Z) - Cross-Lingual Transfer Robustness to Lower-Resource Languages on Adversarial Datasets [4.653113033432781]
Cross-lingual transfer capabilities of Multilingual Language Models (MLLMs) are investigated.
Our research provides valuable insights into cross-lingual transfer and its implications for NLP applications.
arXiv Detail & Related papers (2024-03-29T08:47:15Z) - UltraLink: An Open-Source Knowledge-Enhanced Multilingual Supervised
Fine-tuning Dataset [69.33424532827608]
Open-source large language models (LLMs) have gained significant strength across diverse fields.
In this work, we construct an open-source multilingual supervised fine-tuning dataset.
The resulting UltraLink dataset comprises approximately 1 million samples across five languages.
arXiv Detail & Related papers (2024-02-07T05:05:53Z) - Exposing Cross-Lingual Lexical Knowledge from Multilingual Sentence
Encoders [85.80950708769923]
We probe multilingual language models for the amount of cross-lingual lexical knowledge stored in their parameters, and compare them against the original multilingual LMs.
We also devise a novel method to expose this knowledge by additionally fine-tuning multilingual models.
We report substantial gains on standard benchmarks.
arXiv Detail & Related papers (2022-04-30T13:23:16Z) - Prix-LM: Pretraining for Multilingual Knowledge Base Construction [59.02868906044296]
We propose a unified framework, Prix-LM, for multilingual knowledge construction and completion.
We leverage two types of knowledge, monolingual triples and cross-lingual links, extracted from existing multilingual KBs.
Experiments on standard entity-related tasks, such as link prediction in multiple languages, cross-lingual entity linking and bilingual lexicon induction, demonstrate its effectiveness.
arXiv Detail & Related papers (2021-10-16T02:08:46Z) - Cross-Lingual Transfer in Zero-Shot Cross-Language Entity Linking [19.083300046605252]
Cross-language entity linking grounds mentions in multiple languages to a single-language knowledge base.
We find that the multilingual ability of BERT leads to robust performance in monolingual and multilingual settings.
arXiv Detail & Related papers (2020-10-19T20:08:26Z) - FILTER: An Enhanced Fusion Method for Cross-lingual Language
Understanding [85.29270319872597]
We propose an enhanced fusion method that takes cross-lingual data as input for XLM finetuning.
During inference, the model makes predictions based on the text input in the target language and its translation in the source language.
To tackle this issue, we propose an additional KL-divergence self-teaching loss for model training, based on auto-generated soft pseudo-labels for translated text in the target language.
arXiv Detail & Related papers (2020-09-10T22:42:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.