Pack Together: Entity and Relation Extraction with Levitated Marker
- URL: http://arxiv.org/abs/2109.06067v1
- Date: Mon, 13 Sep 2021 15:38:13 GMT
- Title: Pack Together: Entity and Relation Extraction with Levitated Marker
- Authors: Deming Ye, Yankai Lin, Maosong Sun
- Abstract summary: We propose a novel span representation approach, named Packed Levitated Markers, to consider the dependencies between the spans (pairs) by strategically packing the markers in the encoder.
Our experiments show that our model with packed levitated markers outperforms the sequence labeling model by 0.4%-1.9% F1 on three flat NER tasks, and beats the token concat model on six NER benchmarks.
- Score: 61.232174424421025
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Named Entity Recognition (NER) and Relation Extraction (RE) are the core
sub-tasks for information extraction. Many recent works formulate these two
tasks as the span (pair) classification problem, and thus focus on
investigating how to obtain a better span representation from the pre-trained
encoder. However, a major limitation of existing works is that they ignore the
dependencies between spans (pairs). In this work, we propose a novel span
representation approach, named Packed Levitated Markers, to consider the
dependencies between the spans (pairs) by strategically packing the markers in
the encoder. In particular, we propose a group packing strategy to enable our
model to process massive spans together to consider their dependencies with
limited resources. Furthermore, for those more complicated span pair
classification tasks, we design a subject-oriented packing strategy, which
packs each subject and all its objects into an instance to model the
dependencies between the same-subject span pairs. Our experiments show that our
model with packed levitated markers outperforms the sequence labeling model by
0.4%-1.9% F1 on three flat NER tasks, beats the token concat model on six NER
benchmarks, and obtains a 3.5%-3.6% strict relation F1 improvement with higher
speed over previous SOTA models on ACE04 and ACE05. Code and models are
publicly available at https://github.com/thunlp/PL-Marker.
Related papers
- Mutually Guided Few-shot Learning for Relational Triple Extraction [10.539566491939844]
Mutually Guided Few-shot learning framework for Triple Extraction (MG-FTE)
Our method consists of an entity-guided relation-decoder to classify relations and a proto-decoder to extract entities.
Our method outperforms many state-of-the-art methods by 12.6 F1 score on FewRel 1.0 (single domain) and 20.5 F1 score on FewRel 2.0 (cross-domain)
arXiv Detail & Related papers (2023-06-23T06:15:54Z) - Reference Twice: A Simple and Unified Baseline for Few-Shot Instance Segmentation [103.90033029330527]
Few-Shot Instance (FSIS) requires detecting and segmenting novel classes with limited support examples.
We introduce a unified framework, Reference Twice (RefT), to exploit the relationship between support and query features for FSIS.
arXiv Detail & Related papers (2023-01-03T15:33:48Z) - ReSel: N-ary Relation Extraction from Scientific Text and Tables by
Learning to Retrieve and Select [53.071352033539526]
We study the problem of extracting N-ary relations from scientific articles.
Our proposed method ReSel decomposes this task into a two-stage procedure.
Our experiments on three scientific information extraction datasets show that ReSel outperforms state-of-the-art baselines significantly.
arXiv Detail & Related papers (2022-10-26T02:28:02Z) - SpanProto: A Two-stage Span-based Prototypical Network for Few-shot
Named Entity Recognition [45.012327072558975]
Few-shot Named Entity Recognition (NER) aims to identify named entities with very little annotated data.
We propose a seminal span-based prototypical network (SpanProto) that tackles few-shot NER via a two-stage approach.
In the span extraction stage, we transform the sequential tags into a global boundary matrix, enabling the model to focus on the explicit boundary information.
For mention classification, we leverage prototypical learning to capture the semantic representations for each labeled span and make the model better adapt to novel-class entities.
arXiv Detail & Related papers (2022-10-17T12:59:33Z) - Effective Token Graph Modeling using a Novel Labeling Strategy for
Structured Sentiment Analysis [39.770652220521384]
State-of-the-art model for structured sentiment analysis casts the task as a dependency parsing problem.
Label proportions for span prediction and span relation prediction are imbalanced.
Two nodes in a dependency graph cannot have multiple arcs, therefore some overlapped sentiments cannot be recognized.
arXiv Detail & Related papers (2022-03-21T08:23:03Z) - X2Parser: Cross-Lingual and Cross-Domain Framework for Task-Oriented
Compositional Semantic Parsing [51.81533991497547]
Task-oriented compositional semantic parsing (TCSP) handles complex nested user queries.
We present X2 compared a transferable Cross-lingual and Cross-domain for TCSP.
We propose to predict flattened intents and slots representations separately and cast both prediction tasks into sequence labeling problems.
arXiv Detail & Related papers (2021-06-07T16:40:05Z) - Boosting Span-based Joint Entity and Relation Extraction via Squence
Tagging Mechanism [10.894755638322]
Span-based joint extraction simultaneously conducts named entity recognition (NER) and relation extraction (RE) in text span form.
Recent studies have shown that token labels can convey crucial task-specific information and enrich token semantics.
We pro-pose Sequence Tagging enhanced Span-based Network (STSN), a span-based joint extrac-tion network that is enhanced by token BIO label information.
arXiv Detail & Related papers (2021-05-21T01:10:03Z) - AutoRC: Improving BERT Based Relation Classification Models via
Architecture Search [50.349407334562045]
BERT based relation classification (RC) models have achieved significant improvements over the traditional deep learning models.
No consensus can be reached on what is the optimal architecture.
We design a comprehensive search space for BERT based RC models and employ neural architecture search (NAS) method to automatically discover the design choices.
arXiv Detail & Related papers (2020-09-22T16:55:49Z) - Multi-Label Text Classification using Attention-based Graph Neural
Network [0.0]
A graph attention network-based model is proposed to capture the attentive dependency structure among the labels.
The proposed model achieves similar or better performance compared to the previous state-of-the-art models.
arXiv Detail & Related papers (2020-03-22T17:12:43Z) - AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses [97.50616524350123]
We build dialogue models that are dynamically aware of what utterances or tokens are dull without any feature-engineering.
The first model, MinAvgOut, directly maximizes the diversity score through the output distributions of each batch.
The second model, Label Fine-Tuning (LFT), prepends to the source sequence a label continuously scaled by the diversity score to control the diversity level.
The third model, RL, adopts Reinforcement Learning and treats the diversity score as a reward signal.
arXiv Detail & Related papers (2020-01-15T18:32:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.