Automatic Extraction of Medication Names in Tweets as Named Entity
Recognition
- URL: http://arxiv.org/abs/2111.15641v1
- Date: Tue, 30 Nov 2021 18:25:32 GMT
- Title: Automatic Extraction of Medication Names in Tweets as Named Entity
Recognition
- Authors: Carol Anderson, Bo Liu, Anas Abidin, Hoo-Chang Shin, Virginia Adams
- Abstract summary: Biocreative VII Task 3 focuses on mining this information by recognizing mentions of medications and dietary supplements in tweets.
We approach this task by fine tuning multiple BERT-style language models to perform token-level classification.
Our best system consists of five Megatron-BERT-345M models and achieves a strict F1 score of 0.764 on unseen test data.
- Score: 3.7462395049372894
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Social media posts contain potentially valuable information about medical
conditions and health-related behavior. Biocreative VII Task 3 focuses on
mining this information by recognizing mentions of medications and dietary
supplements in tweets. We approach this task by fine tuning multiple BERT-style
language models to perform token-level classification, and combining them into
ensembles to generate final predictions. Our best system consists of five
Megatron-BERT-345M models and achieves a strict F1 score of 0.764 on unseen
test data.
Related papers
- Towards Evaluating and Building Versatile Large Language Models for Medicine [57.49547766838095]
We present MedS-Bench, a benchmark designed to evaluate the performance of large language models (LLMs) in clinical contexts.
MedS-Bench spans 11 high-level clinical tasks, including clinical report summarization, treatment recommendations, diagnosis, named entity recognition, and medical concept explanation.
MedS-Ins comprises 58 medically oriented language corpora, totaling 13.5 million samples across 122 tasks.
arXiv Detail & Related papers (2024-08-22T17:01:34Z) - Hybrid X-Linker: Automated Data Generation and Extreme Multi-label Ranking for Biomedical Entity Linking [45.16091578348614]
State-of-the-art deep learning entity linking methods rely on extensive human-labelled data.
Current datasets are limited in size, leading to inadequate coverage of biomedical concepts.
We propose to automatically generate data to create large-scale training datasets.
arXiv Detail & Related papers (2024-07-08T18:04:22Z) - BiomedRAG: A Retrieval Augmented Large Language Model for Biomedicine [19.861178160437827]
Large Language Models (LLMs) have swiftly emerged as vital resources for different applications in the biomedical and healthcare domains.
textscBiomedRAG attains superior performance across 5 biomedical NLP tasks.
textscBiomedRAG outperforms other triple extraction systems with micro-F1 scores of 81.42 and 88.83 on GIT and ChemProt corpora, respectively.
arXiv Detail & Related papers (2024-05-01T12:01:39Z) - ThangDLU at #SMM4H 2024: Encoder-decoder models for classifying text data on social disorders in children and adolescents [49.00494558898933]
This paper describes our participation in Task 3 and Task 5 of the #SMM4H (Social Media Mining for Health) 2024 Workshop.
Task 3 is a multi-class classification task centered on tweets discussing the impact of outdoor environments on symptoms of social anxiety.
Task 5 involves a binary classification task focusing on tweets reporting medical disorders in children.
We applied transfer learning from pre-trained encoder-decoder models such as BART-base and T5-small to identify the labels of a set of given tweets.
arXiv Detail & Related papers (2024-04-30T17:06:20Z) - PMC-LLaMA: Towards Building Open-source Language Models for Medicine [62.39105735933138]
Large Language Models (LLMs) have showcased remarkable capabilities in natural language understanding.
LLMs struggle in domains that require precision, such as medical applications, due to their lack of domain-specific knowledge.
We describe the procedure for building a powerful, open-source language model specifically designed for medicine applications, termed as PMC-LLaMA.
arXiv Detail & Related papers (2023-04-27T18:29:05Z) - BCH-NLP at BioCreative VII Track 3: medications detection in tweets
using transformer networks and multi-task learning [9.176393163624002]
We implement a multi-task learning model that is jointly trained on text classification and sequence labelling.
Our best system run achieved a strict F1 of 80.4, ranking first and more than 10 points higher than the average score of all participants.
arXiv Detail & Related papers (2021-11-26T19:22:51Z) - Extraction of Medication Names from Twitter Using Augmentation and an
Ensemble of Language Models [55.44979919361194]
The BioCreative VII Track 3 challenge focused on the identification of medication names in Twitter user timelines.
For our submission to this challenge, we expanded the available training data by using several data augmentation techniques.
The augmented data was then used to fine-tune an ensemble of language models that had been pre-trained on general-domain Twitter content.
arXiv Detail & Related papers (2021-11-12T11:18:46Z) - A PubMedBERT-based Classifier with Data Augmentation Strategy for
Detecting Medication Mentions in Tweets [2.539568419434224]
Twitter publishes a large number of user-generated text (tweets) on a daily basis.
entity recognition (NER) presents some special challenges for tweet data.
In this paper, we explore a PubMedBERT-based classifier trained with a combination of multiple data augmentation approaches.
Our method achieved an F1 score of 0.762, which is substantially higher than the mean of all submissions.
arXiv Detail & Related papers (2021-11-03T14:29:24Z) - CLIP: A Dataset for Extracting Action Items for Physicians from Hospital
Discharge Notes [17.107315598110183]
We create a dataset of clinical action items annotated over MIMIC-III, the largest publicly available dataset of real clinical notes.
This dataset, which we call CLIP, is annotated by physicians and covers documents representing 100K sentences.
We describe the task of extracting the action items from these documents as multi-aspect extractive summarization, with each aspect representing a type of action to be taken.
arXiv Detail & Related papers (2021-06-04T14:49:02Z) - Students Need More Attention: BERT-based AttentionModel for Small Data
with Application to AutomaticPatient Message Triage [65.7062363323781]
We propose a novel framework based on BioBERT (Bidirectional Representations from Transformers forBiomedical TextMining)
We introduce Label Embeddings for Self-Attention in each layer of BERT, which we call LESA-BERT, and (ii) by distilling LESA-BERT to smaller variants, we aim to reduce overfitting and model size when working on small datasets.
As an application, our framework is utilized to build a model for patient portal message triage that classifies the urgency of a message into three categories: non-urgent, medium and urgent.
arXiv Detail & Related papers (2020-06-22T03:39:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.