A hybrid deep-learning approach for complex biochemical named entity
recognition
- URL: http://arxiv.org/abs/2012.10824v1
- Date: Sun, 20 Dec 2020 01:30:07 GMT
- Title: A hybrid deep-learning approach for complex biochemical named entity
recognition
- Authors: Jian Liu, Lei Gao, Sujie Guo, Rui Ding, Xin Huang, Long Ye, Qinghua
Meng, Asef Nazari and Dhananjay Thiruvady
- Abstract summary: Named entity recognition (NER) of chemicals and drugs is a critical domain of information extraction in biochemical research.
Here, we propose a hybrid deep learning approach to improve the recognition accuracy of NER.
- Score: 9.657827522380712
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Named entity recognition (NER) of chemicals and drugs is a critical domain of
information extraction in biochemical research. NER provides support for text
mining in biochemical reactions, including entity relation extraction,
attribute extraction, and metabolic response relationship extraction. However,
the existence of complex naming characteristics in the biomedical field, such
as polysemy and special characters, make the NER task very challenging. Here,
we propose a hybrid deep learning approach to improve the recognition accuracy
of NER. Specifically, our approach applies the Bidirectional Encoder
Representations from Transformers (BERT) model to extract the underlying
features of the text, learns a representation of the context of the text
through Bi-directional Long Short-Term Memory (BILSTM), and incorporates the
multi-head attention (MHATT) mechanism to extract chapter-level features. In
this approach, the MHATT mechanism aims to improve the recognition accuracy of
abbreviations to efficiently deal with the problem of inconsistency in
full-text labels. Moreover, conditional random field (CRF) is used to label
sequence tags because this probabilistic method does not need strict
independence assumptions and can accommodate arbitrary context information. The
experimental evaluation on a publicly-available dataset shows that the proposed
hybrid approach achieves the best recognition performance; in particular, it
substantially improves performance in recognizing abbreviations, polysemes, and
low-frequency entities, compared with the state-of-the-art approaches. For
instance, compared with the recognition accuracies for low-frequency entities
produced by the BILSTM-CRF algorithm, those produced by the hybrid approach on
two entity datasets (MULTIPLE and IDENTIFIER) have been increased by 80% and
21.69%, respectively.
Related papers
- Unlocking Potential Binders: Multimodal Pretraining DEL-Fusion for Denoising DNA-Encoded Libraries [51.72836644350993]
Multimodal Pretraining DEL-Fusion model (MPDF)
We develop pretraining tasks applying contrastive objectives between different compound representations and their text descriptions.
We propose a novel DEL-fusion framework that amalgamates compound information at the atomic, submolecular, and molecular levels.
arXiv Detail & Related papers (2024-09-07T17:32:21Z) - Prototype Learning Guided Hybrid Network for Breast Tumor Segmentation in DCE-MRI [58.809276442508256]
We propose a hybrid network via the combination of convolution neural network (CNN) and transformer layers.
The experimental results on private and public DCE-MRI datasets demonstrate that the proposed hybrid network superior performance than the state-of-the-art methods.
arXiv Detail & Related papers (2024-08-11T15:46:00Z) - BioMNER: A Dataset for Biomedical Method Entity Recognition [25.403593761614424]
We propose a novel dataset for biomedical method entity recognition.
We employ an automated BioMethod entity recognition and information retrieval system to assist human annotation.
Our empirical findings reveal that the large parameter counts of language models surprisingly inhibit the effective assimilation of entity extraction patterns.
arXiv Detail & Related papers (2024-06-28T16:34:24Z) - Spatial Semantic Recurrent Mining for Referring Image Segmentation [63.34997546393106]
We propose Stextsuperscript2RM to achieve high-quality cross-modality fusion.
It follows a working strategy of trilogy: distributing language feature, spatial semantic recurrent coparsing, and parsed-semantic balancing.
Our proposed method performs favorably against other state-of-the-art algorithms.
arXiv Detail & Related papers (2024-05-15T00:17:48Z) - Joint Multimodal Transformer for Emotion Recognition in the Wild [49.735299182004404]
Multimodal emotion recognition (MMER) systems typically outperform unimodal systems.
This paper proposes an MMER method that relies on a joint multimodal transformer (JMT) for fusion with key-based cross-attention.
arXiv Detail & Related papers (2024-03-15T17:23:38Z) - EMBRE: Entity-aware Masking for Biomedical Relation Extraction [12.821610050561256]
We introduce the Entity-aware Masking for Biomedical Relation Extraction (EMBRE) method for relation extraction.
Specifically, we integrate entity knowledge into a deep neural network by pretraining the backbone model with an entity masking objective.
arXiv Detail & Related papers (2024-01-15T18:12:01Z) - Multi-level biomedical NER through multi-granularity embeddings and
enhanced labeling [3.8599767910528917]
This paper proposes a hybrid approach that integrates the strengths of multiple models.
BERT provides contextualized word embeddings, a pre-trained multi-channel CNN for character-level information capture, and following by a BiLSTM + CRF for sequence labelling and modelling dependencies between the words in the text.
We evaluate our model on the benchmark i2b2/2010 dataset, achieving an F1-score of 90.11.
arXiv Detail & Related papers (2023-12-24T21:45:36Z) - BERM: Training the Balanced and Extractable Representation for Matching
to Improve Generalization Ability of Dense Retrieval [54.66399120084227]
We propose a novel method to improve the generalization of dense retrieval via capturing matching signal called BERM.
Dense retrieval has shown promise in the first-stage retrieval process when trained on in-domain labeled datasets.
arXiv Detail & Related papers (2023-05-18T15:43:09Z) - BERT-GT: Cross-sentence n-ary relation extraction with BERT and Graph
Transformer [7.262905275276971]
We propose a novel architecture that combines Bidirectional Representations from Transformers with Graph Transformer (BERT-GT)
Unlike the original Transformer architecture, which utilizes the whole sentence(s) to calculate the attention of the current token, the neighbor-attention mechanism in our method calculates its attention utilizing only its neighbor tokens.
Our results show improvements of 5.44% and 3.89% in accuracy and F1-measure over the state-of-the-art on n-proteinary and chemical-proteinary datasets.
arXiv Detail & Related papers (2021-01-11T19:34:55Z) - Biomedical named entity recognition using BERT in the machine reading
comprehension framework [16.320249089801884]
We propose a new method to implement biomedical named entity recognition (BioNER)
Instead of treating the BioNER task as a sequence labeling problem, we formulate it as a machine reading comprehension problem.
Our method achieves state-of-the-art (SOTA) performance on the BC4CHEMD, BC5CDR-Chem, BC5CDR-Disease, NCBI-Disease, BC2GM and JNLPBA datasets.
arXiv Detail & Related papers (2020-09-03T10:10:20Z) - Text Recognition in Real Scenarios with a Few Labeled Samples [55.07859517380136]
Scene text recognition (STR) is still a hot research topic in computer vision field.
This paper proposes a few-shot adversarial sequence domain adaptation (FASDA) approach to build sequence adaptation.
Our approach can maximize the character-level confusion between the source domain and the target domain.
arXiv Detail & Related papers (2020-06-22T13:03:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.