Investigation of BERT Model on Biomedical Relation Extraction Based on
Revised Fine-tuning Mechanism
- URL: http://arxiv.org/abs/2011.00398v1
- Date: Sun, 1 Nov 2020 01:47:16 GMT
- Title: Investigation of BERT Model on Biomedical Relation Extraction Based on
Revised Fine-tuning Mechanism
- Authors: Peng Su, K. Vijay-Shanker
- Abstract summary: We will investigate the method of utilizing the entire layer in the fine-tuning process of BERT model.
In addition, further analysis indicates that the key knowledge about the relations can be learned from the last layer of BERT model.
- Score: 2.8881198461098894
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With the explosive growth of biomedical literature, designing automatic tools
to extract information from the literature has great significance in biomedical
research. Recently, transformer-based BERT models adapted to the biomedical
domain have produced leading results. However, all the existing BERT models for
relation classification only utilize partial knowledge from the last layer. In
this paper, we will investigate the method of utilizing the entire layer in the
fine-tuning process of BERT model. To the best of our knowledge, we are the
first to explore this method. The experimental results illustrate that our
method improves the BERT model performance and outperforms the state-of-the-art
methods on three benchmark datasets for different relation extraction tasks. In
addition, further analysis indicates that the key knowledge about the relations
can be learned from the last layer of BERT model.
Related papers
- Deep Learning for Medical Text Processing: BERT Model Fine-Tuning and Comparative Study [4.416456130207115]
This paper proposes a medical literature summary generation method based on the BERT model to address the challenges brought by the current explosion of medical information.
We develop an efficient summary generation system that can quickly extract key information from medical literature and generate coherent, accurate summaries.
arXiv Detail & Related papers (2024-10-28T07:17:45Z) - BioBERT-based Deep Learning and Merged ChemProt-DrugProt for Enhanced Biomedical Relation Extraction [2.524192238862961]
Our approach integrates the ChemProt and DrugProt datasets using a novel merging strategy.
The study highlights the potential of automated information extraction in biomedical research and clinical practice.
arXiv Detail & Related papers (2024-05-28T21:34:01Z) - MS-MANO: Enabling Hand Pose Tracking with Biomechanical Constraints [50.61346764110482]
We integrate a musculoskeletal system with a learnable parametric hand model, MANO, to create MS-MANO.
This model emulates the dynamics of muscles and tendons to drive the skeletal system, imposing physiologically realistic constraints on the resulting torque trajectories.
We also propose a simulation-in-the-loop pose refinement framework, BioPR, that refines the initial estimated pose through a multi-layer perceptron network.
arXiv Detail & Related papers (2024-04-16T02:18:18Z) - Research on the Application of Deep Learning-based BERT Model in
Sentiment Analysis [8.504422968998506]
This paper explores the application of deep learning techniques, particularly focusing on BERT models, in sentiment analysis.
It elucidates the application effects and optimization strategies of BERT models in sentiment analysis, supported by experimental validation.
The experimental findings indicate that BERT models exhibit robust performance in sentiment analysis tasks, with notable enhancements post fine-tuning.
arXiv Detail & Related papers (2024-03-13T03:31:26Z) - Multi-level biomedical NER through multi-granularity embeddings and
enhanced labeling [3.8599767910528917]
This paper proposes a hybrid approach that integrates the strengths of multiple models.
BERT provides contextualized word embeddings, a pre-trained multi-channel CNN for character-level information capture, and following by a BiLSTM + CRF for sequence labelling and modelling dependencies between the words in the text.
We evaluate our model on the benchmark i2b2/2010 dataset, achieving an F1-score of 90.11.
arXiv Detail & Related papers (2023-12-24T21:45:36Z) - Diversifying Knowledge Enhancement of Biomedical Language Models using
Adapter Modules and Knowledge Graphs [54.223394825528665]
We develop an approach that uses lightweight adapter modules to inject structured biomedical knowledge into pre-trained language models.
We use two large KGs, the biomedical knowledge system UMLS and the novel biochemical OntoChem, with two prominent biomedical PLMs, PubMedBERT and BioLinkBERT.
We show that our methodology leads to performance improvements in several instances while keeping requirements in computing power low.
arXiv Detail & Related papers (2023-12-21T14:26:57Z) - Improving Biomedical Entity Linking with Retrieval-enhanced Learning [53.24726622142558]
$k$NN-BioEL provides a BioEL model with the ability to reference similar instances from the entire training corpus as clues for prediction.
We show that $k$NN-BioEL outperforms state-of-the-art baselines on several datasets.
arXiv Detail & Related papers (2023-12-15T14:04:23Z) - Fine-Tuning Large Neural Language Models for Biomedical Natural Language
Processing [55.52858954615655]
We conduct a systematic study on fine-tuning stability in biomedical NLP.
We show that finetuning performance may be sensitive to pretraining settings, especially in low-resource domains.
We show that these techniques can substantially improve fine-tuning performance for lowresource biomedical NLP applications.
arXiv Detail & Related papers (2021-12-15T04:20:35Z) - Discovering Drug-Target Interaction Knowledge from Biomedical Literature [107.98712673387031]
The Interaction between Drugs and Targets (DTI) in human body plays a crucial role in biomedical science and applications.
As millions of papers come out every year in the biomedical domain, automatically discovering DTI knowledge from literature becomes an urgent demand in the industry.
We explore the first end-to-end solution for this task by using generative approaches.
We regard the DTI triplets as a sequence and use a Transformer-based model to directly generate them without using the detailed annotations of entities and relations.
arXiv Detail & Related papers (2021-09-27T17:00:14Z) - Evaluating Biomedical BERT Models for Vocabulary Alignment at Scale in
the UMLS Metathesaurus [8.961270657070942]
The current UMLS (Unified Medical Language System) Metathesaurus construction process is expensive and error-prone.
Recent advances in Natural Language Processing have achieved state-of-the-art (SOTA) performance on downstream tasks.
We aim to validate if approaches using the BERT models can actually outperform the existing approaches for predicting synonymy in the UMLS Metathesaurus.
arXiv Detail & Related papers (2021-09-14T16:52:16Z) - Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation [84.64004917951547]
Fine-tuning pre-trained language models like BERT has become an effective way in NLP.
In this paper, we improve the fine-tuning of BERT with two effective mechanisms: self-ensemble and self-distillation.
arXiv Detail & Related papers (2020-02-24T16:17:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.