Chem-FINESE: Validating Fine-Grained Few-shot Entity Extraction through Text Reconstruction
- URL: http://arxiv.org/abs/2401.10189v4
- Date: Wed, 29 May 2024 18:24:15 GMT
- Title: Chem-FINESE: Validating Fine-Grained Few-shot Entity Extraction through Text Reconstruction
- Authors: Qingyun Wang, Zixuan Zhang, Hongxiang Li, Xuan Liu, Jiawei Han, Huimin Zhao, Heng Ji,
- Abstract summary: Fine-grained few-shot entity extraction in the chemical domain faces two unique challenges.
Chem-FINESE has two components: a seq2seq entity extractor and a seq2seq self-validation module.
Our newly proposed framework has contributed up to 8.26% and 6.84% absolute F1-score gains respectively.
- Score: 68.76468780148734
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fine-grained few-shot entity extraction in the chemical domain faces two unique challenges. First, compared with entity extraction tasks in the general domain, sentences from chemical papers usually contain more entities. Moreover, entity extraction models usually have difficulty extracting entities of long-tailed types. In this paper, we propose Chem-FINESE, a novel sequence-to-sequence (seq2seq) based few-shot entity extraction approach, to address these two challenges. Our Chem-FINESE has two components: a seq2seq entity extractor to extract named entities from the input sentence and a seq2seq self-validation module to reconstruct the original input sentence from extracted entities. Inspired by the fact that a good entity extraction system needs to extract entities faithfully, our new self-validation module leverages entity extraction results to reconstruct the original input sentence. Besides, we design a new contrastive loss to reduce excessive copying during the extraction process. Finally, we release ChemNER+, a new fine-grained chemical entity extraction dataset that is annotated by domain experts with the ChemNER schema. Experiments in few-shot settings with both ChemNER+ and CHEMET datasets show that our newly proposed framework has contributed up to 8.26% and 6.84% absolute F1-score gains respectively.
Related papers
- MARE: Multi-Aspect Rationale Extractor on Unsupervised Rationale Extraction [10.998983921416533]
Unsupervised rationale extraction aims to extract text snippets to support model predictions without explicit rationale annotation.
Previous works often encode each aspect independently, which may limit their ability to capture meaningful internal correlations between aspects.
We propose a Multi-Aspect Rationale Extractor (MARE) to explain and predict multiple aspects simultaneously.
arXiv Detail & Related papers (2024-10-04T15:52:29Z) - Extract-and-Abstract: Unifying Extractive and Abstractive Summarization within Single Encoder-Decoder Framework [24.97672212363703]
We propose ExtAbs, which jointly and seamlessly performs Extractive and Abstractive summarization tasks within single encoder-decoder model.
In ExtAbs, the vanilla encoder is augmented to extract salients, and the vanilla decoder is modified with the proposed saliency mask to generate summaries.
Experiments show that ExtAbs can achieve superior performance than baselines on the extractive task and performs comparable, or even better than the vanilla models on the abstractive task.
arXiv Detail & Related papers (2024-09-18T09:21:25Z) - Learning to Extract Structured Entities Using Language Models [52.281701191329]
Recent advances in machine learning have significantly impacted the field of information extraction.
We reformulate the task to be entity-centric, enabling the use of diverse metrics.
We contribute to the field by introducing Structured Entity Extraction and proposing the Approximate Entity Set OverlaP metric.
arXiv Detail & Related papers (2024-02-06T22:15:09Z) - S2F-NER: Exploring Sequence-to-Forest Generation for Complex Entity
Recognition [47.714230389689064]
We propose a novel Sequence-to-Forest generation paradigm, S2F-NER, which can directly extract entities in sentence via a Forest decoder.
Specifically, our model generate each path of each tree in forest autoregressively, where the maximum depth of each tree is three.
Based on this novel paradigm, our model can elegantly mitigate the exposure bias problem and keep the simplicity of Seq2Seq.
arXiv Detail & Related papers (2023-10-29T09:09:10Z) - End-to-End Models for Chemical-Protein Interaction Extraction: Better
Tokenization and Span-Based Pipeline Strategies [1.782718930156674]
We employ a span-based pipeline approach to produce a new state-of-the-art E2ERE performance on the ChemProt dataset.
Our results indicate that a straightforward fine-grained tokenization scheme helps span-based approaches excel in E2ERE.
arXiv Detail & Related papers (2023-04-03T20:20:22Z) - Summarization Programs: Interpretable Abstractive Summarization with
Neural Modular Trees [89.60269205320431]
Current abstractive summarization models either suffer from a lack of clear interpretability or provide incomplete rationales.
We propose the Summarization Program (SP), an interpretable modular framework consisting of an (ordered) list of binary trees.
A Summarization Program contains one root node per summary sentence, and a distinct tree connects each summary sentence to the document sentences.
arXiv Detail & Related papers (2022-09-21T16:50:22Z) - Nested Named Entity Recognition as Latent Lexicalized Constituency
Parsing [29.705133932275892]
Recently, (Fu et al, 2021) adapt a span-based constituency to tackle nested NER.
In this work, we resort to more expressive structures, lexicalized constituency trees in which constituents are annotated by headwords.
We leverage the Eisner-Satta algorithm to perform partial marginalization and inference efficiently.
arXiv Detail & Related papers (2022-03-09T12:02:59Z) - Document-level Entity-based Extraction as Template Generation [13.110360825201044]
We propose a generative framework for two document-level EE tasks: role-filler entity extraction (REE) and relation extraction (RE)
We first formulate them as a template generation problem, allowing models to efficiently capture cross-entity dependencies.
A novel cross-attention guided copy mechanism, TopK Copy, is incorporated into a pre-trained sequence-to-sequence model to enhance the capabilities of identifying key information.
arXiv Detail & Related papers (2021-09-10T14:18:22Z) - MIMOSA: Multi-constraint Molecule Sampling for Molecule Optimization [51.00815310242277]
generative models and reinforcement learning approaches made initial success, but still face difficulties in simultaneously optimizing multiple drug properties.
We propose the MultI-constraint MOlecule SAmpling (MIMOSA) approach, a sampling framework to use input molecule as an initial guess and sample molecules from the target distribution.
arXiv Detail & Related papers (2020-10-05T20:18:42Z) - At Which Level Should We Extract? An Empirical Analysis on Extractive
Document Summarization [110.54963847339775]
We show that unnecessity and redundancy issues exist when extracting full sentences.
We propose extracting sub-sentential units based on the constituency parsing tree.
arXiv Detail & Related papers (2020-04-06T13:35:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.