Chem-FINESE: Validating Fine-Grained Few-shot Entity Extraction through Text Reconstruction
- URL: http://arxiv.org/abs/2401.10189v4
- Date: Wed, 29 May 2024 18:24:15 GMT
- Title: Chem-FINESE: Validating Fine-Grained Few-shot Entity Extraction through Text Reconstruction
- Authors: Qingyun Wang, Zixuan Zhang, Hongxiang Li, Xuan Liu, Jiawei Han, Huimin Zhao, Heng Ji,
- Abstract summary: Fine-grained few-shot entity extraction in the chemical domain faces two unique challenges.
Chem-FINESE has two components: a seq2seq entity extractor and a seq2seq self-validation module.
Our newly proposed framework has contributed up to 8.26% and 6.84% absolute F1-score gains respectively.
- Score: 68.76468780148734
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Fine-grained few-shot entity extraction in the chemical domain faces two unique challenges. First, compared with entity extraction tasks in the general domain, sentences from chemical papers usually contain more entities. Moreover, entity extraction models usually have difficulty extracting entities of long-tailed types. In this paper, we propose Chem-FINESE, a novel sequence-to-sequence (seq2seq) based few-shot entity extraction approach, to address these two challenges. Our Chem-FINESE has two components: a seq2seq entity extractor to extract named entities from the input sentence and a seq2seq self-validation module to reconstruct the original input sentence from extracted entities. Inspired by the fact that a good entity extraction system needs to extract entities faithfully, our new self-validation module leverages entity extraction results to reconstruct the original input sentence. Besides, we design a new contrastive loss to reduce excessive copying during the extraction process. Finally, we release ChemNER+, a new fine-grained chemical entity extraction dataset that is annotated by domain experts with the ChemNER schema. Experiments in few-shot settings with both ChemNER+ and CHEMET datasets show that our newly proposed framework has contributed up to 8.26% and 6.84% absolute F1-score gains respectively.
Related papers
- S2F-NER: Exploring Sequence-to-Forest Generation for Complex Entity
Recognition [47.714230389689064]
We propose a novel Sequence-to-Forest generation paradigm, S2F-NER, which can directly extract entities in sentence via a Forest decoder.
Specifically, our model generate each path of each tree in forest autoregressively, where the maximum depth of each tree is three.
Based on this novel paradigm, our model can elegantly mitigate the exposure bias problem and keep the simplicity of Seq2Seq.
arXiv Detail & Related papers (2023-10-29T09:09:10Z) - End-to-End Models for Chemical-Protein Interaction Extraction: Better
Tokenization and Span-Based Pipeline Strategies [1.782718930156674]
We employ a span-based pipeline approach to produce a new state-of-the-art E2ERE performance on the ChemProt dataset.
Our results indicate that a straightforward fine-grained tokenization scheme helps span-based approaches excel in E2ERE.
arXiv Detail & Related papers (2023-04-03T20:20:22Z) - Summarization Programs: Interpretable Abstractive Summarization with
Neural Modular Trees [89.60269205320431]
Current abstractive summarization models either suffer from a lack of clear interpretability or provide incomplete rationales.
We propose the Summarization Program (SP), an interpretable modular framework consisting of an (ordered) list of binary trees.
A Summarization Program contains one root node per summary sentence, and a distinct tree connects each summary sentence to the document sentences.
arXiv Detail & Related papers (2022-09-21T16:50:22Z) - Nested Named Entity Recognition as Latent Lexicalized Constituency
Parsing [29.705133932275892]
Recently, (Fu et al, 2021) adapt a span-based constituency to tackle nested NER.
In this work, we resort to more expressive structures, lexicalized constituency trees in which constituents are annotated by headwords.
We leverage the Eisner-Satta algorithm to perform partial marginalization and inference efficiently.
arXiv Detail & Related papers (2022-03-09T12:02:59Z) - DYLE: Dynamic Latent Extraction for Abstractive Long-Input Summarization [19.693324720440017]
Transformer-based models have achieved state-of-the-art performance on short text summarization.
We present a new approach for long-input summarization: Dynamic Latent Extraction for Abstractive Summarization.
Our model significantly outperforms the current state-of-the-art, including a 6.21 ROUGE-2 improvement on GovReport and a 2.13 ROUGE-1 improvement on QMSum.
arXiv Detail & Related papers (2021-10-15T15:53:32Z) - Document-level Entity-based Extraction as Template Generation [13.110360825201044]
We propose a generative framework for two document-level EE tasks: role-filler entity extraction (REE) and relation extraction (RE)
We first formulate them as a template generation problem, allowing models to efficiently capture cross-entity dependencies.
A novel cross-attention guided copy mechanism, TopK Copy, is incorporated into a pre-trained sequence-to-sequence model to enhance the capabilities of identifying key information.
arXiv Detail & Related papers (2021-09-10T14:18:22Z) - MIMOSA: Multi-constraint Molecule Sampling for Molecule Optimization [51.00815310242277]
generative models and reinforcement learning approaches made initial success, but still face difficulties in simultaneously optimizing multiple drug properties.
We propose the MultI-constraint MOlecule SAmpling (MIMOSA) approach, a sampling framework to use input molecule as an initial guess and sample molecules from the target distribution.
arXiv Detail & Related papers (2020-10-05T20:18:42Z) - Contrastive Triple Extraction with Generative Transformer [72.21467482853232]
We introduce a novel model, contrastive triple extraction with a generative transformer.
Specifically, we introduce a single shared transformer module for encoder-decoder-based generation.
To generate faithful results, we propose a novel triplet contrastive training object.
arXiv Detail & Related papers (2020-09-14T05:29:24Z) - At Which Level Should We Extract? An Empirical Analysis on Extractive
Document Summarization [110.54963847339775]
We show that unnecessity and redundancy issues exist when extracting full sentences.
We propose extracting sub-sentential units based on the constituency parsing tree.
arXiv Detail & Related papers (2020-04-06T13:35:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.