Fast and Effective Biomedical Entity Linking Using a Dual Encoder
- URL: http://arxiv.org/abs/2103.05028v1
- Date: Mon, 8 Mar 2021 19:32:28 GMT
- Title: Fast and Effective Biomedical Entity Linking Using a Dual Encoder
- Authors: Rajarshi Bhowmik and Karl Stratos and Gerard de Melo
- Abstract summary: We propose a BERT-based dual encoder model that resolves multiple mentions in a document in one shot.
We show that our proposed model is multiple times faster than existing BERT-based models while being competitive in accuracy for biomedical entity linking.
- Score: 48.86736921025866
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Biomedical entity linking is the task of identifying mentions of biomedical
concepts in text documents and mapping them to canonical entities in a target
thesaurus. Recent advancements in entity linking using BERT-based models follow
a retrieve and rerank paradigm, where the candidate entities are first selected
using a retriever model, and then the retrieved candidates are ranked by a
reranker model. While this paradigm produces state-of-the-art results, they are
slow both at training and test time as they can process only one mention at a
time. To mitigate these issues, we propose a BERT-based dual encoder model that
resolves multiple mentions in a document in one shot. We show that our proposed
model is multiple times faster than existing BERT-based models while being
competitive in accuracy for biomedical entity linking. Additionally, we modify
our dual encoder model for end-to-end biomedical entity linking that performs
both mention span detection and entity disambiguation and out-performs two
recently proposed models.
Related papers
- Entity Disambiguation via Fusion Entity Decoding [68.77265315142296]
We propose an encoder-decoder model to disambiguate entities with more detailed entity descriptions.
We observe +1.5% improvements in end-to-end entity linking in the GERBIL benchmark compared with EntQA.
arXiv Detail & Related papers (2024-04-02T04:27:54Z) - Biomedical Entity Linking as Multiple Choice Question Answering [48.74212158495695]
We present BioELQA, a novel model that treats Biomedical Entity Linking as Multiple Choice Question Answering.
We first obtains candidate entities with a fast retriever, jointly presents the mention and candidate entities to a generator, and then outputs the predicted symbol associated with its chosen entity.
To improve generalization for long-tailed entities, we retrieve similar labeled training instances as clues and the input with retrieved instances for the generator.
arXiv Detail & Related papers (2024-02-23T08:40:38Z) - Single-Stage Visual Relationship Learning using Conditional Queries [60.90880759475021]
TraCQ is a new formulation for scene graph generation that avoids the multi-task learning problem and the entity pair distribution.
We employ a DETR-based encoder-decoder conditional queries to significantly reduce the entity label space as well.
Experimental results show that TraCQ not only outperforms existing single-stage scene graph generation methods, it also beats many state-of-the-art two-stage methods on the Visual Genome dataset.
arXiv Detail & Related papers (2023-06-09T06:02:01Z) - Partial Annotation Learning for Biomedical Entity Recognition [0.19336815376402716]
We show that partial annotation learning methods can effectively learn from biomedical corpora with missing entity annotations.
Our proposed model outperforms alternatives and, specifically, the PubMedBERT tagger by 38% in F1-score under high missing entity rates.
arXiv Detail & Related papers (2023-05-22T15:18:38Z) - Nested Named Entity Recognition from Medical Texts: An Adaptive Shared
Network Architecture with Attentive CRF [53.55504611255664]
We propose a novel method, referred to as ASAC, to solve the dilemma caused by the nested phenomenon.
The proposed method contains two key modules: the adaptive shared (AS) part and the attentive conditional random field (ACRF) module.
Our model could learn better entity representations by capturing the implicit distinctions and relationships between different categories of entities.
arXiv Detail & Related papers (2022-11-09T09:23:56Z) - ReFinED: An Efficient Zero-shot-capable Approach to End-to-End Entity
Linking [5.382800665115746]
ReFinED is an efficient end-to-end entity linking model.
It performs mention detection, fine-grained entity typing, and entity disambiguation for all mentions within a document in a single forward pass.
It surpasses state-of-the-art performance on standard entity linking datasets by an average of 3.7 F1.
arXiv Detail & Related papers (2022-07-08T19:20:42Z) - Knowledge-Rich Self-Supervised Entity Linking [58.838404666183656]
Knowledge-RIch Self-Supervision ($tt KRISSBERT$) is a universal entity linker for four million UMLS entities.
Our approach subsumes zero-shot and few-shot methods, and can easily incorporate entity descriptions and gold mention labels if available.
Without using any labeled information, our method produces $tt KRISSBERT$, a universal entity linker for four million UMLS entities.
arXiv Detail & Related papers (2021-12-15T05:05:12Z) - BERT might be Overkill: A Tiny but Effective Biomedical Entity Linker
based on Residual Convolutional Neural Networks [41.528797439272175]
We propose an efficient convolutional neural network with residual connections for biomedical entity linking.
Our model achieves comparable or even better linking accuracy than the state-of-the-art BERT-based models.
arXiv Detail & Related papers (2021-09-06T04:25:47Z) - A Lightweight Neural Model for Biomedical Entity Linking [1.8047694351309205]
We propose a lightweight neural method for biomedical entity linking.
Our method uses a simple alignment layer with attention mechanisms to capture the variations between mention and entity names.
Our model is competitive with previous work on standard evaluation benchmarks.
arXiv Detail & Related papers (2020-12-16T10:34:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.