Injecting Categorical Labels and Syntactic Information into Biomedical
NER
- URL: http://arxiv.org/abs/2311.03113v1
- Date: Mon, 6 Nov 2023 14:03:59 GMT
- Title: Injecting Categorical Labels and Syntactic Information into Biomedical
NER
- Authors: Sumam Francis, Marie-Francine Moens
- Abstract summary: We present a simple approach to improve biomedical named entity recognition (NER) by injecting categorical labels and Part-of-speech (POS) information into the model.
Experiments on three benchmark datasets show that incorporating categorical label information with syntactic context is quite useful and outperforms baseline BERT-based models.
- Score: 28.91836510067532
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a simple approach to improve biomedical named entity recognition
(NER) by injecting categorical labels and Part-of-speech (POS) information into
the model. We use two approaches, in the first approach, we first train a
sequence-level classifier to classify the sentences into categories to obtain
the sentence-level tags (categorical labels). The sequence classifier is
modeled as an entailment problem by modifying the labels as a natural language
template. This helps to improve the accuracy of the classifier. Further, this
label information is injected into the NER model. In this paper, we demonstrate
effective ways to represent and inject these labels and POS attributes into the
NER model. In the second approach, we jointly learn the categorical labels and
NER labels. Here we also inject the POS tags into the model to increase the
syntactic context of the model. Experiments on three benchmark datasets show
that incorporating categorical label information with syntactic context is
quite useful and outperforms baseline BERT-based models.
Related papers
- Posterior Label Smoothing for Node Classification [2.737276507021477]
We propose a simple yet effective label smoothing for the transductive node classification task.
We design the soft label to encapsulate the local context of the target node through the neighborhood label distribution.
In the following analysis, we find that incorporating global label statistics in posterior computation is the key to the success of label smoothing.
arXiv Detail & Related papers (2024-06-01T11:59:49Z) - Description-Enhanced Label Embedding Contrastive Learning for Text
Classification [65.01077813330559]
Self-Supervised Learning (SSL) in model learning process and design a novel self-supervised Relation of Relation (R2) classification task.
Relation of Relation Learning Network (R2-Net) for text classification, in which text classification and R2 classification are treated as optimization targets.
external knowledge from WordNet to obtain multi-aspect descriptions for label semantic learning.
arXiv Detail & Related papers (2023-06-15T02:19:34Z) - Bridging the Gap between Model Explanations in Partially Annotated
Multi-label Classification [85.76130799062379]
We study how false negative labels affect the model's explanation.
We propose to boost the attribution scores of the model trained with partial labels to make its explanation resemble that of the model trained with full labels.
arXiv Detail & Related papers (2023-04-04T14:00:59Z) - Ground Truth Inference for Weakly Supervised Entity Matching [76.6732856489872]
We propose a simple but powerful labeling model for weak supervision tasks.
We then tailor the labeling model specifically to the task of entity matching.
We show that our labeling model results in a 9% higher F1 score on average than the best existing method.
arXiv Detail & Related papers (2022-11-13T17:57:07Z) - Enhancing Label Consistency on Document-level Named Entity Recognition [19.249781091058605]
Named entity recognition (NER) is a fundamental part of extracting information from documents in biomedical applications.
We present our method, ConNER, which enhances the label dependency of modifier (e.g., adjectives and prepositions) to achieve higher label agreement.
The effectiveness of our method is demonstrated on four popular biomedical NER datasets.
arXiv Detail & Related papers (2022-10-24T04:45:17Z) - Optimizing Bi-Encoder for Named Entity Recognition via Contrastive
Learning [80.36076044023581]
We present an efficient bi-encoder framework for named entity recognition (NER)
We frame NER as a metric learning problem that maximizes the similarity between the vector representations of an entity mention and its type.
A major challenge to this bi-encoder formulation for NER lies in separating non-entity spans from entity mentions.
arXiv Detail & Related papers (2022-08-30T23:19:04Z) - Binary Classification with Positive Labeling Sources [71.37692084951355]
We propose WEAPO, a simple yet competitive WS method for producing training labels without negative labeling sources.
We show WEAPO achieves the highest averaged performance on 10 benchmark datasets.
arXiv Detail & Related papers (2022-08-02T19:32:08Z) - Label Semantics for Few Shot Named Entity Recognition [68.01364012546402]
We study the problem of few shot learning for named entity recognition.
We leverage the semantic information in the names of the labels as a way of giving the model additional signal and enriched priors.
Our model learns to match the representations of named entities computed by the first encoder with label representations computed by the second encoder.
arXiv Detail & Related papers (2022-03-16T23:21:05Z) - A Unified Generative Adversarial Network Training via Self-Labeling and
Self-Attention [38.31735499785227]
We propose a novel GAN training scheme that can handle any level of labeling in a unified manner.
Our scheme introduces a form of artificial labeling that can incorporate manually defined labels, when available.
We evaluate our approach on CIFAR-10, STL-10 and SVHN, and show that both self-labeling and self-attention consistently improve the quality of generated data.
arXiv Detail & Related papers (2021-06-18T04:40:26Z) - Improving Pretrained Models for Zero-shot Multi-label Text
Classification through Reinforced Label Hierarchy Reasoning [18.531022315325583]
Exploiting label hierarchies has become a promising approach to tackling the zero-shot multi-label text classification problem.
We propose a Reinforced Label Hierarchy Reasoning (RLHR) approach to encourage interdependence among labels in the hierarchies during training.
arXiv Detail & Related papers (2021-04-04T19:14:09Z) - Label Confusion Learning to Enhance Text Classification Models [3.0251266104313643]
Label Confusion Model (LCM) learns label confusion to capture semantic overlap among labels.
LCM can generate a better label distribution to replace the original one-hot label vector.
experiments on five text classification benchmark datasets reveal the effectiveness of LCM for several widely used deep learning classification models.
arXiv Detail & Related papers (2020-12-09T11:34:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.