Biomedical Entity Representations with Synonym Marginalization
- URL: http://arxiv.org/abs/2005.00239v1
- Date: Fri, 1 May 2020 06:20:36 GMT
- Title: Biomedical Entity Representations with Synonym Marginalization
- Authors: Mujeen Sung, Hwisang Jeon, Jinhyuk Lee, Jaewoo Kang
- Abstract summary: We focus on learning representations of biomedical entities solely based on the synonyms of entities.
Our model-based candidates are iteratively updated to contain more difficult negative samples as our model evolves.
Our model BioSyn consistently outperforms previous state-of-the-art models almost reaching the upper bound on each dataset.
- Score: 23.051019207472027
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Biomedical named entities often play important roles in many biomedical text
mining tools. However, due to the incompleteness of provided synonyms and
numerous variations in their surface forms, normalization of biomedical
entities is very challenging. In this paper, we focus on learning
representations of biomedical entities solely based on the synonyms of
entities. To learn from the incomplete synonyms, we use a model-based candidate
selection and maximize the marginal likelihood of the synonyms present in top
candidates. Our model-based candidates are iteratively updated to contain more
difficult negative samples as our model evolves. In this way, we avoid the
explicit pre-selection of negative samples from more than 400K candidates. On
four biomedical entity normalization datasets having three different entity
types (disease, chemical, adverse reaction), our model BioSyn consistently
outperforms previous state-of-the-art models almost reaching the upper bound on
each dataset.
Related papers
- BioT5+: Towards Generalized Biological Understanding with IUPAC Integration and Multi-task Tuning [77.90250740041411]
This paper introduces BioT5+, an extension of the BioT5 framework, tailored to enhance biological research and drug discovery.
BioT5+ incorporates several novel features: integration of IUPAC names for molecular understanding, inclusion of extensive bio-text and molecule data from sources like bioRxiv and PubChem, the multi-task instruction tuning for generality across tasks, and a numerical tokenization technique for improved processing of numerical data.
arXiv Detail & Related papers (2024-02-27T12:43:09Z) - Biomedical Entity Linking as Multiple Choice Question Answering [48.74212158495695]
We present BioELQA, a novel model that treats Biomedical Entity Linking as Multiple Choice Question Answering.
We first obtains candidate entities with a fast retriever, jointly presents the mention and candidate entities to a generator, and then outputs the predicted symbol associated with its chosen entity.
To improve generalization for long-tailed entities, we retrieve similar labeled training instances as clues and the input with retrieved instances for the generator.
arXiv Detail & Related papers (2024-02-23T08:40:38Z) - Biomedical Language Models are Robust to Sub-optimal Tokenization [30.175714262031253]
Most modern biomedical language models (LMs) are pre-trained using standard domain-specific tokenizers.
We find that pre-training a biomedical LM using a more accurate biomedical tokenizer does not improve the entity representation quality of a language model.
arXiv Detail & Related papers (2023-06-30T13:35:24Z) - BiomedGPT: A Generalist Vision-Language Foundation Model for Diverse Biomedical Tasks [68.39821375903591]
Generalist AI holds the potential to address limitations due to its versatility in interpreting different data types.
Here, we propose BiomedGPT, the first open-source and lightweight vision-language foundation model.
arXiv Detail & Related papers (2023-05-26T17:14:43Z) - Biomedical Named Entity Recognition via Dictionary-based Synonym
Generalization [51.89486520806639]
We propose a novel Synonym Generalization (SynGen) framework that recognizes the biomedical concepts contained in the input text using span-based predictions.
We extensively evaluate our approach on a wide range of benchmarks and the results verify that SynGen outperforms previous dictionary-based models by notable margins.
arXiv Detail & Related papers (2023-05-22T14:36:32Z) - BioGPT: Generative Pre-trained Transformer for Biomedical Text
Generation and Mining [140.61707108174247]
We propose BioGPT, a domain-specific generative Transformer language model pre-trained on large scale biomedical literature.
We get 44.98%, 38.42% and 40.76% F1 score on BC5CDR, KD-DTI and DDI end-to-end relation extraction tasks respectively, and 78.2% accuracy on PubMedQA.
arXiv Detail & Related papers (2022-10-19T07:17:39Z) - Biomedical Interpretable Entity Representations [40.6095537182194]
Pre-trained language models induce dense entity representations that offer strong performance on entity-centric NLP tasks.
This can be a barrier to model uptake in important domains such as biomedicine.
We create a new entity type system and training set from a large corpus of biomedical texts.
arXiv Detail & Related papers (2021-06-17T13:52:10Z) - Fast and Effective Biomedical Entity Linking Using a Dual Encoder [48.86736921025866]
We propose a BERT-based dual encoder model that resolves multiple mentions in a document in one shot.
We show that our proposed model is multiple times faster than existing BERT-based models while being competitive in accuracy for biomedical entity linking.
arXiv Detail & Related papers (2021-03-08T19:32:28Z) - How Do Your Biomedical Named Entity Models Generalize to Novel Entities? [17.83980569600546]
We analyze the three types of recognition abilities of BioNER models: memorization, synonym generalization, and concept generalization.
We find that (1) BioNER models are overestimated in terms of their generalization ability, and (2) they tend to exploit dataset biases, which hinders the models' abilities to generalize.
Our method consistently improves the generalizability of the state-of-the-art (SOTA) models on five benchmark datasets, allowing them to better perform on unseen entity mentions.
arXiv Detail & Related papers (2021-01-01T04:13:42Z) - A Lightweight Neural Model for Biomedical Entity Linking [1.8047694351309205]
We propose a lightweight neural method for biomedical entity linking.
Our method uses a simple alignment layer with attention mechanisms to capture the variations between mention and entity names.
Our model is competitive with previous work on standard evaluation benchmarks.
arXiv Detail & Related papers (2020-12-16T10:34:37Z) - BioALBERT: A Simple and Effective Pre-trained Language Model for
Biomedical Named Entity Recognition [9.05154470433578]
Existing BioNER approaches often neglect these issues and directly adopt the state-of-the-art (SOTA) models.
We propose biomedical ALBERT, an effective domain-specific language model trained on large-scale biomedical corpora.
arXiv Detail & Related papers (2020-09-19T12:58:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.