Solving the Right Problem is Key for Translational NLP: A Case Study in
UMLS Vocabulary Insertion
- URL: http://arxiv.org/abs/2311.15106v1
- Date: Sat, 25 Nov 2023 19:35:53 GMT
- Title: Solving the Right Problem is Key for Translational NLP: A Case Study in
UMLS Vocabulary Insertion
- Authors: Bernal Jimenez Gutierrez, Yuqing Mao, Vinh Nguyen, Kin Wah Fung, Yu
Su, Olivier Bodenreider
- Abstract summary: We study the case of UMLS vocabulary insertion, an important real-world task in which hundreds of thousands of new terms are added to the UMLS.
We introduce a new formulation for UMLS vocabulary insertion which mirrors the real-world task.
We also propose an effective rule-enhanced biomedical language model which enables important new model behavior.
- Score: 12.855898113768998
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As the immense opportunities enabled by large language models become more
apparent, NLP systems will be increasingly expected to excel in real-world
settings. However, in many instances, powerful models alone will not yield
translational NLP solutions, especially if the formulated problem is not well
aligned with the real-world task. In this work, we study the case of UMLS
vocabulary insertion, an important real-world task in which hundreds of
thousands of new terms, referred to as atoms, are added to the UMLS, one of the
most comprehensive open-source biomedical knowledge bases. Previous work aimed
to develop an automated NLP system to make this time-consuming, costly, and
error-prone task more efficient. Nevertheless, practical progress in this
direction has been difficult to achieve due to a problem formulation and
evaluation gap between research output and the real-world task. In order to
address this gap, we introduce a new formulation for UMLS vocabulary insertion
which mirrors the real-world task, datasets which faithfully represent it and
several strong baselines we developed through re-purposing existing solutions.
Additionally, we propose an effective rule-enhanced biomedical language model
which enables important new model behavior, outperforms all strong baselines
and provides measurable qualitative improvements to editors who carry out the
UVI task. We hope this case study provides insight into the considerable
importance of problem formulation for the success of translational NLP
solutions.
Related papers
- BloomWise: Enhancing Problem-Solving capabilities of Large Language Models using Bloom's-Taxonomy-Inspired Prompts [59.83547898874152]
We introduce BloomWise, a new prompting technique, inspired by Bloom's taxonomy, to improve the performance of Large Language Models (LLMs)
The decision regarding the need to employ more sophisticated cognitive skills is based on self-evaluation performed by the LLM.
In extensive experiments across 4 popular math reasoning datasets, we have demonstrated the effectiveness of our proposed approach.
arXiv Detail & Related papers (2024-10-05T09:27:52Z) - Easy Problems That LLMs Get Wrong [0.0]
We introduce a comprehensive Linguistic Benchmark designed to evaluate the limitations of Large Language Models (LLMs)
Through a series of straightforward questions, it uncovers the significant limitations of well-regarded models to perform tasks that humans manage with ease.
arXiv Detail & Related papers (2024-05-30T02:09:51Z) - ALMol: Aligned Language-Molecule Translation LLMs through Offline Preference Contrastive Optimisation [2.296475290901356]
We focus on machine language-molecule translation and deploy a novel training approach called contrastive preference optimisation.
Our results demonstrate that our models achieve up to a 32% improvement compared to counterpart models.
arXiv Detail & Related papers (2024-05-14T13:59:24Z) - Scalable Language Model with Generalized Continual Learning [58.700439919096155]
The Joint Adaptive Re-ization (JARe) is integrated with Dynamic Task-related Knowledge Retrieval (DTKR) to enable adaptive adjustment of language models based on specific downstream tasks.
Our method demonstrates state-of-the-art performance on diverse backbones and benchmarks, achieving effective continual learning in both full-set and few-shot scenarios with minimal forgetting.
arXiv Detail & Related papers (2024-04-11T04:22:15Z) - Corex: Pushing the Boundaries of Complex Reasoning through Multi-Model Collaboration [83.4031923134958]
Corex is a suite of novel general-purpose strategies that transform Large Language Models into autonomous agents.
Inspired by human behaviors, Corex is constituted by diverse collaboration paradigms including Debate, Review, and Retrieve modes.
We demonstrate that orchestrating multiple LLMs to work in concert yields substantially better performance compared to existing methods.
arXiv Detail & Related papers (2023-09-30T07:11:39Z) - Simultaneous Machine Translation with Large Language Models [51.470478122113356]
We investigate the possibility of applying Large Language Models to SimulMT tasks.
We conducted experiments using the textttLlama2-7b-chat model on nine different languages from the MUST-C dataset.
The results show that LLM outperforms dedicated MT models in terms of BLEU and LAAL metrics.
arXiv Detail & Related papers (2023-09-13T04:06:47Z) - Automatically Correcting Large Language Models: Surveying the landscape
of diverse self-correction strategies [104.32199881187607]
Large language models (LLMs) have demonstrated remarkable performance across a wide array of NLP tasks.
A promising approach to rectify these flaws is self-correction, where the LLM itself is prompted or guided to fix problems in its own output.
This paper presents a comprehensive review of this emerging class of techniques.
arXiv Detail & Related papers (2023-08-06T18:38:52Z) - A Survey of Large Language Models [81.06947636926638]
Language modeling has been widely studied for language understanding and generation in the past two decades.
Recently, pre-trained language models (PLMs) have been proposed by pre-training Transformer models over large-scale corpora.
To discriminate the difference in parameter scale, the research community has coined the term large language models (LLM) for the PLMs of significant size.
arXiv Detail & Related papers (2023-03-31T17:28:46Z) - Model-Agnostic Meta-Learning for Natural Language Understanding Tasks in
Finance [1.863067234952186]
We investigate model-agnostic meta-learning algorithm(MAML) in low-resource financial NLU tasks.
Our models achieve the state-of-the-art performance according to the experimental results.
arXiv Detail & Related papers (2023-03-06T02:24:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.