Type-supervised sequence labeling based on the heterogeneous star graph
for named entity recognition
- URL: http://arxiv.org/abs/2210.10240v2
- Date: Fri, 21 Oct 2022 13:21:50 GMT
- Title: Type-supervised sequence labeling based on the heterogeneous star graph
for named entity recognition
- Authors: Xueru Wen, Changjiang Zhou, Haotian Tang, Luguang Liang, Yu Jiang,
Hong Qi
- Abstract summary: The representation learning of the heterogeneous star graph containing text nodes and type nodes is investigated in this paper.
The model performs the type-supervised sequence labeling after updating nodes in the graph.
Experiments on public NER datasets reveal the effectiveness of our model in extracting both flat and nested entities.
- Score: 6.25916397918329
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Named entity recognition is a fundamental task in natural language
processing, identifying the span and category of entities in unstructured
texts. The traditional sequence labeling methodology ignores the nested
entities, i.e. entities included in other entity mentions. Many approaches
attempt to address this scenario, most of which rely on complex structures or
have high computation complexity. The representation learning of the
heterogeneous star graph containing text nodes and type nodes is investigated
in this paper. In addition, we revise the graph attention mechanism into a
hybrid form to address its unreasonableness in specific topologies. The model
performs the type-supervised sequence labeling after updating nodes in the
graph. The annotation scheme is an extension of the single-layer sequence
labeling and is able to cope with the vast majority of nested entities.
Extensive experiments on public NER datasets reveal the effectiveness of our
model in extracting both flat and nested entities. The method achieved
state-of-the-art performance on both flat and nested datasets. The significant
improvement in accuracy reflects the superiority of the multi-layer labeling
strategy.
Related papers
- Scribbles for All: Benchmarking Scribble Supervised Segmentation Across Datasets [51.74296438621836]
We introduce Scribbles for All, a label and training data generation algorithm for semantic segmentation trained on scribble labels.
The main limitation of scribbles as source for weak supervision is the lack of challenging datasets for scribble segmentation.
Scribbles for All provides scribble labels for several popular segmentation datasets and provides an algorithm to automatically generate scribble labels for any dataset with dense annotations.
arXiv Detail & Related papers (2024-08-22T15:29:08Z) - Hypergraph based Understanding for Document Semantic Entity Recognition [65.84258776834524]
We build a novel hypergraph attention document semantic entity recognition framework, HGA, which uses hypergraph attention to focus on entity boundaries and entity categories at the same time.
Our results on FUNSD, CORD, XFUNDIE show that our method can effectively improve the performance of semantic entity recognition tasks.
arXiv Detail & Related papers (2024-07-09T14:35:49Z) - Multi-label Node Classification On Graph-Structured Data [7.892731722253387]
Graph Neural Networks (GNNs) have shown state-of-the-art improvements in node classification tasks on graphs.
A more general and realistic scenario in which each node could have multiple labels has so far received little attention.
We collect and release three real-world biological datasets and develop a multi-label graph generator.
arXiv Detail & Related papers (2023-04-20T15:34:20Z) - GrannGAN: Graph annotation generative adversarial networks [72.66289932625742]
We consider the problem of modelling high-dimensional distributions and generating new examples of data with complex relational feature structure coherent with a graph skeleton.
The model we propose tackles the problem of generating the data features constrained by the specific graph structure of each data point by splitting the task into two phases.
In the first it models the distribution of features associated with the nodes of the given graph, in the second it complements the edge features conditionally on the node features.
arXiv Detail & Related papers (2022-12-01T11:49:07Z) - SpanProto: A Two-stage Span-based Prototypical Network for Few-shot
Named Entity Recognition [45.012327072558975]
Few-shot Named Entity Recognition (NER) aims to identify named entities with very little annotated data.
We propose a seminal span-based prototypical network (SpanProto) that tackles few-shot NER via a two-stage approach.
In the span extraction stage, we transform the sequential tags into a global boundary matrix, enabling the model to focus on the explicit boundary information.
For mention classification, we leverage prototypical learning to capture the semantic representations for each labeled span and make the model better adapt to novel-class entities.
arXiv Detail & Related papers (2022-10-17T12:59:33Z) - Trigger-GNN: A Trigger-Based Graph Neural Network for Nested Named
Entity Recognition [5.9049664765234295]
We propose a trigger-based graph neural network (Trigger-GNN) to leverage the nested NER.
It obtains the complementary annotation embeddings through entity trigger encoding and semantic matching.
It helps the model to learn and generalize more efficiently and cost-effectively.
arXiv Detail & Related papers (2022-04-12T04:15:39Z) - Multi-task Self-distillation for Graph-based Semi-Supervised Learning [6.277952154365413]
We propose a multi-task self-distillation framework that injects self-supervised learning and self-distillation into graph convolutional networks.
First, we formulate a self-supervision pipeline based on pre-text tasks to capture different levels of similarities in graphs.
Second, self-distillation uses soft labels of the model itself as additional supervision.
arXiv Detail & Related papers (2021-12-02T12:43:41Z) - Learning the Implicit Semantic Representation on Graph-Structured Data [57.670106959061634]
Existing representation learning methods in graph convolutional networks are mainly designed by describing the neighborhood of each node as a perceptual whole.
We propose a Semantic Graph Convolutional Networks (SGCN) that explores the implicit semantics by learning latent semantic-paths in graphs.
arXiv Detail & Related papers (2021-01-16T16:18:43Z) - Fine-Grained Named Entity Typing over Distantly Supervised Data Based on
Refined Representations [16.30478830298353]
Fine-Grained Named Entity Typing (FG-NET) is a key component in Natural Language Processing (NLP)
We propose an edge-weighted attentive graph convolution network that refines the noisy mention representations by attending over corpus-level contextual clues prior to the end classification.
Experimental evaluation shows that the proposed model outperforms the existing research by a relative score of upto 10.2% and 8.3% for macro f1 and micro f1 respectively.
arXiv Detail & Related papers (2020-04-07T17:26:36Z) - Weakly-Supervised Salient Object Detection via Scribble Annotations [54.40518383782725]
We propose a weakly-supervised salient object detection model to learn saliency from scribble labels.
We present a new metric, termed saliency structure measure, to measure the structure alignment of the predicted saliency maps.
Our method not only outperforms existing weakly-supervised/unsupervised methods, but also is on par with several fully-supervised state-of-the-art models.
arXiv Detail & Related papers (2020-03-17T12:59:50Z) - Graph Inference Learning for Semi-supervised Classification [50.55765399527556]
We propose a Graph Inference Learning framework to boost the performance of semi-supervised node classification.
For learning the inference process, we introduce meta-optimization on structure relations from training nodes to validation nodes.
Comprehensive evaluations on four benchmark datasets demonstrate the superiority of our proposed GIL when compared against state-of-the-art methods.
arXiv Detail & Related papers (2020-01-17T02:52:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.