Embeddings of Label Components for Sequence Labeling: A Case Study of
Fine-grained Named Entity Recognition
- URL: http://arxiv.org/abs/2006.01372v2
- Date: Thu, 4 Jun 2020 14:01:18 GMT
- Title: Embeddings of Label Components for Sequence Labeling: A Case Study of
Fine-grained Named Entity Recognition
- Authors: Takuma Kato, Kaori Abe, Hiroki Ouchi, Shumpei Miyawaki, Jun Suzuki,
Kentaro Inui
- Abstract summary: We propose to integrate label component information as embeddings into models.
We demonstrate that the proposed method improves performance, especially for instances with low-frequency labels.
- Score: 41.60109880213463
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In general, the labels used in sequence labeling consist of different types
of elements. For example, IOB-format entity labels, such as B-Person and
I-Person, can be decomposed into span (B and I) and type information (Person).
However, while most sequence labeling models do not consider such label
components, the shared components across labels, such as Person, can be
beneficial for label prediction. In this work, we propose to integrate label
component information as embeddings into models. Through experiments on English
and Japanese fine-grained named entity recognition, we demonstrate that the
proposed method improves performance, especially for instances with
low-frequency labels.
Related papers
- Exploiting Conjugate Label Information for Multi-Instance Partial-Label Learning [61.00359941983515]
Multi-instance partial-label learning (MIPL) addresses scenarios where each training sample is represented as a multi-instance bag associated with a candidate label set containing one true label and several false positives.
ELIMIPL exploits the conjugate label information to improve the disambiguation performance.
arXiv Detail & Related papers (2024-08-26T15:49:31Z) - Substituting Data Annotation with Balanced Updates and Collective Loss
in Multi-label Text Classification [19.592985329023733]
Multi-label text classification (MLTC) is the task of assigning multiple labels to a given text.
We study the MLTC problem in annotation-free and scarce-annotation settings in which the magnitude of available supervision signals is linear to the number of labels.
Our method follows three steps, (1) mapping input text into a set of preliminary label likelihoods by natural language inference using a pre-trained language model, (2) calculating a signed label dependency graph by label descriptions, and (3) updating the preliminary label likelihoods with message passing along the label dependency graph.
arXiv Detail & Related papers (2023-09-24T04:12:52Z) - Bridging the Gap between Model Explanations in Partially Annotated
Multi-label Classification [85.76130799062379]
We study how false negative labels affect the model's explanation.
We propose to boost the attribution scores of the model trained with partial labels to make its explanation resemble that of the model trained with full labels.
arXiv Detail & Related papers (2023-04-04T14:00:59Z) - Towards Few-shot Entity Recognition in Document Images: A Label-aware
Sequence-to-Sequence Framework [28.898240725099782]
We build an entity recognition model requiring only a few shots of annotated document images.
We develop a novel label-aware seq2seq framework, LASER.
Experiments on two benchmark datasets demonstrate the superiority of LASER under the few-shot setting.
arXiv Detail & Related papers (2022-03-30T18:30:42Z) - Semantic-Aware Representation Blending for Multi-Label Image Recognition
with Partial Labels [86.17081952197788]
We propose to blend category-specific representation across different images to transfer information of known labels to complement unknown labels.
Experiments on the MS-COCO, Visual Genome, Pascal VOC 2007 datasets show that the proposed SARB framework obtains superior performance over current leading competitors.
arXiv Detail & Related papers (2022-03-04T07:56:16Z) - Structured Semantic Transfer for Multi-Label Recognition with Partial
Labels [85.6967666661044]
We propose a structured semantic transfer (SST) framework that enables training multi-label recognition models with partial labels.
The framework consists of two complementary transfer modules that explore within-image and cross-image semantic correlations.
Experiments on the Microsoft COCO, Visual Genome and Pascal VOC datasets show that the proposed SST framework obtains superior performance over current state-of-the-art algorithms.
arXiv Detail & Related papers (2021-12-21T02:15:01Z) - Evaluating Multi-label Classifiers with Noisy Labels [0.7868449549351487]
In the real world, it is more common to deal with noisy datasets than clean datasets.
We present a Context-Based Multi-Label-Classifier (CbMLC) that effectively handles noisy labels.
We show CbMLC yields substantial improvements over the previous methods in most cases.
arXiv Detail & Related papers (2021-02-16T19:50:52Z) - A Study on the Autoregressive and non-Autoregressive Multi-label
Learning [77.11075863067131]
We propose a self-attention based variational encoder-model to extract the label-label and label-feature dependencies jointly.
Our model can therefore be used to predict all labels in parallel while still including both label-label and label-feature dependencies.
arXiv Detail & Related papers (2020-12-03T05:41:44Z) - iobes: A Library for Span-Level Processing [11.112281331309939]
iobe is used for parsing, converting, and processing spans represented as token level decisions.
In this paper, we introduce our open remedy, iobe.
arXiv Detail & Related papers (2020-10-09T05:03:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.