A New Entity Extraction Method Based on Machine Reading Comprehension
- URL: http://arxiv.org/abs/2108.06444v1
- Date: Sat, 14 Aug 2021 02:11:41 GMT
- Title: A New Entity Extraction Method Based on Machine Reading Comprehension
- Authors: Xiaobo Jiang, Kun He, Jiajun He and Guangyu Yan
- Abstract summary: This paper presents an effective MRC-based entity extraction model-MRC-I2DP.
It uses the proposed gated attention-attracting mechanism to adjust the restoration of each part of the text pair.
It also uses the proposed 2D probability coding module, TALU function and mask mechanism to strengthen the detection of all possible targets of the target.
- Score: 4.92025078254413
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Entity extraction is a key technology for obtaining information from massive
texts in natural language processing. The further interaction between them does
not meet the standards of human reading comprehension, thus limiting the
understanding of the model, and also the omission or misjudgment of the answer
(ie the target entity) due to the reasoning question. An effective MRC-based
entity extraction model-MRC-I2DP, which uses the proposed gated
attention-attracting mechanism to adjust the restoration of each part of the
text pair, creating problems and thinking for multi-level interactive attention
calculations to increase the target entity It also uses the proposed 2D
probability coding module, TALU function and mask mechanism to strengthen the
detection of all possible targets of the target, thereby improving the
probability and accuracy of prediction. Experiments have proved that MRC-I2DP
represents an overall state-of-the-art model in 7 from the scientific and
public domains, achieving a performance improvement of 2.1% ~ 10.4% compared to
the model model in F1.
Related papers
- The Buffer Mechanism for Multi-Step Information Reasoning in Language Models [52.77133661679439]
Investigating internal reasoning mechanisms of large language models can help us design better model architectures and training strategies.
In this study, we constructed a symbolic dataset to investigate the mechanisms by which Transformer models employ vertical thinking strategy.
We proposed a random matrix-based algorithm to enhance the model's reasoning ability, resulting in a 75% reduction in the training time required for the GPT-2 model.
arXiv Detail & Related papers (2024-05-24T07:41:26Z) - Enhancing Pre-Trained Generative Language Models with Question Attended Span Extraction on Machine Reading Comprehension [6.602323571343169]
Integrated during the fine-tuning phase of pre-trained generative language models (PLMs), QASE significantly enhances their performance.
The efficacy of the QASE module has been rigorously tested across various datasets.
arXiv Detail & Related papers (2024-04-27T19:42:51Z) - A Novel Hybrid Feature Importance and Feature Interaction Detection
Framework for Predictive Optimization in Industry 4.0 Applications [1.0870564199697297]
This paper proposes a novel hybrid framework that combines the feature importance detector - local interpretable model-agnostic explanations (LIME) and the feature interaction detector - neural interaction detection (NID)
The experimental outcomes reveal an augmentation of up to 9.56% in the R2 score, and a diminution of up to 24.05% in the root mean square error.
arXiv Detail & Related papers (2024-03-04T13:22:53Z) - IMUGPT 2.0: Language-Based Cross Modality Transfer for Sensor-Based
Human Activity Recognition [0.19791587637442667]
Cross modality transfer approaches convert existing datasets from a source modality, such as video, to a target modality (IMU)
We introduce two new extensions for IMUGPT that enhance its use for practical HAR application scenarios.
We demonstrate that our diversity metrics can reduce the effort needed for the generation of virtual IMU data by at least 50%.
arXiv Detail & Related papers (2024-02-01T22:37:33Z) - Bidirectional Trained Tree-Structured Decoder for Handwritten
Mathematical Expression Recognition [51.66383337087724]
The Handwritten Mathematical Expression Recognition (HMER) task is a critical branch in the field of OCR.
Recent studies have demonstrated that incorporating bidirectional context information significantly improves the performance of HMER models.
We propose the Mirror-Flipped Symbol Layout Tree (MF-SLT) and Bidirectional Asynchronous Training (BAT) structure.
arXiv Detail & Related papers (2023-12-31T09:24:21Z) - When Parameter-efficient Tuning Meets General-purpose Vision-language
Models [65.19127815275307]
PETAL revolutionizes the training process by requiring only 0.5% of the total parameters, achieved through a unique mode approximation technique.
Our experiments reveal that PETAL not only outperforms current state-of-the-art methods in most scenarios but also surpasses full fine-tuning models in effectiveness.
arXiv Detail & Related papers (2023-12-16T17:13:08Z) - Efficient Model-Free Exploration in Low-Rank MDPs [76.87340323826945]
Low-Rank Markov Decision Processes offer a simple, yet expressive framework for RL with function approximation.
Existing algorithms are either (1) computationally intractable, or (2) reliant upon restrictive statistical assumptions.
We propose the first provably sample-efficient algorithm for exploration in Low-Rank MDPs.
arXiv Detail & Related papers (2023-07-08T15:41:48Z) - Clinical Concept and Relation Extraction Using Prompt-based Machine
Reading Comprehension [38.79665143111312]
We formulate both clinical concept extraction and relation extraction using a unified prompt-based machine reading comprehension architecture.
We compare our MRC models with existing deep learning models for concept extraction and end-to-end relation extraction.
We evaluate the transfer learning ability of the proposed MRC models in a cross-institution setting.
arXiv Detail & Related papers (2023-03-14T22:37:31Z) - USER: Unified Semantic Enhancement with Momentum Contrast for Image-Text
Retrieval [115.28586222748478]
Image-Text Retrieval (ITR) aims at searching for the target instances that are semantically relevant to the given query from the other modality.
Existing approaches typically suffer from two major limitations.
arXiv Detail & Related papers (2023-01-17T12:42:58Z) - MoEfication: Conditional Computation of Transformer Models for Efficient
Inference [66.56994436947441]
Transformer-based pre-trained language models can achieve superior performance on most NLP tasks due to large parameter capacity, but also lead to huge computation cost.
We explore to accelerate large-model inference by conditional computation based on the sparse activation phenomenon.
We propose to transform a large model into its mixture-of-experts (MoE) version with equal model size, namely MoEfication.
arXiv Detail & Related papers (2021-10-05T02:14:38Z) - PSD2 Explainable AI Model for Credit Scoring [0.0]
The aim of this project is to develop and test advanced analytical methods to improve the prediction accuracy of Credit Risk Models.
The project focuses on applying an explainable machine learning model to bank-related databases.
arXiv Detail & Related papers (2020-11-20T12:12:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.