None Class Ranking Loss for Document-Level Relation Extraction
- URL: http://arxiv.org/abs/2205.00476v2
- Date: Tue, 3 May 2022 05:27:54 GMT
- Title: None Class Ranking Loss for Document-Level Relation Extraction
- Authors: Yang Zhou and Wee Sun Lee
- Abstract summary: Document-level relation extraction (RE) aims at extracting relations among entities expressed across multiple sentences.
In a typical document, most entity pairs do not express any pre-defined relation and are labeled as "none" or "no relation"
- Score: 22.173080823450498
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Document-level relation extraction (RE) aims at extracting relations among
entities expressed across multiple sentences, which can be viewed as a
multi-label classification problem. In a typical document, most entity pairs do
not express any pre-defined relation and are labeled as "none" or "no
relation". For good document-level RE performance, it is crucial to distinguish
such none class instances (entity pairs) from those of pre-defined classes
(relations). However, most existing methods only estimate the probability of
pre-defined relations independently without considering the probability of "no
relation". This ignores the context of entity pairs and the label correlations
between the none class and pre-defined classes, leading to sub-optimal
predictions. To address this problem, we propose a new multi-label loss that
encourages large margins of label confidence scores between each pre-defined
class and the none class, which enables captured label correlations and
context-dependent thresholding for label prediction. To gain further robustness
against positive-negative imbalance and mislabeled data that could appear in
real-world RE datasets, we propose a margin regularization and a margin
shifting technique. Experimental results demonstrate that our method
significantly outperforms existing multi-label losses for document-level RE and
works well in other multi-label tasks such as emotion classification when none
class instances are available for training.
Related papers
- Multi-Label Classification for Implicit Discourse Relation Recognition [10.280148603465697]
We explore various multi-label classification frameworks to handle implicit discourse relation recognition.
We show that multi-label classification methods don't depress performance for single-label prediction.
arXiv Detail & Related papers (2024-06-06T19:37:25Z) - Generating Unbiased Pseudo-labels via a Theoretically Guaranteed
Chebyshev Constraint to Unify Semi-supervised Classification and Regression [57.17120203327993]
threshold-to-pseudo label process (T2L) in classification uses confidence to determine the quality of label.
In nature, regression also requires unbiased methods to generate high-quality labels.
We propose a theoretically guaranteed constraint for generating unbiased labels based on Chebyshev's inequality.
arXiv Detail & Related papers (2023-11-03T08:39:35Z) - JointMatch: A Unified Approach for Diverse and Collaborative
Pseudo-Labeling to Semi-Supervised Text Classification [65.268245109828]
Semi-supervised text classification (SSTC) has gained increasing attention due to its ability to leverage unlabeled data.
Existing approaches based on pseudo-labeling suffer from the issues of pseudo-label bias and error accumulation.
We propose JointMatch, a holistic approach for SSTC that addresses these challenges by unifying ideas from recent semi-supervised learning.
arXiv Detail & Related papers (2023-10-23T05:43:35Z) - Dist-PU: Positive-Unlabeled Learning from a Label Distribution
Perspective [89.5370481649529]
We propose a label distribution perspective for PU learning in this paper.
Motivated by this, we propose to pursue the label distribution consistency between predicted and ground-truth label distributions.
Experiments on three benchmark datasets validate the effectiveness of the proposed method.
arXiv Detail & Related papers (2022-12-06T07:38:29Z) - An Effective Approach for Multi-label Classification with Missing Labels [8.470008570115146]
We propose a pseudo-label based approach to reduce the cost of annotation without bringing additional complexity to the classification networks.
By designing a novel loss function, we are able to relax the requirement that each instance must contain at least one positive label.
We show that our method can handle the imbalance between positive labels and negative labels, while still outperforming existing missing-label learning approaches.
arXiv Detail & Related papers (2022-10-24T23:13:57Z) - A Unified Positive-Unlabeled Learning Framework for Document-Level
Relation Extraction with Different Levels of Labeling [5.367772036988716]
Document-level relation extraction (RE) aims to identify relations between entities across multiple sentences.
We propose a unified positive-unlabeled learning framework - shift and squared ranking loss.
Our method achieves an improvement of about 14 F1 points relative to the previous baseline with incomplete labeling.
arXiv Detail & Related papers (2022-10-17T02:54:49Z) - Learning with Proper Partial Labels [87.65718705642819]
Partial-label learning is a kind of weakly-supervised learning with inexact labels.
We show that this proper partial-label learning framework includes many previous partial-label learning settings.
We then derive a unified unbiased estimator of the classification risk.
arXiv Detail & Related papers (2021-12-23T01:37:03Z) - A Theory-Driven Self-Labeling Refinement Method for Contrastive
Representation Learning [111.05365744744437]
Unsupervised contrastive learning labels crops of the same image as positives, and other image crops as negatives.
In this work, we first prove that for contrastive learning, inaccurate label assignment heavily impairs its generalization for semantic instance discrimination.
Inspired by this theory, we propose a novel self-labeling refinement approach for contrastive learning.
arXiv Detail & Related papers (2021-06-28T14:24:52Z) - A Flexible Class of Dependence-aware Multi-Label Loss Functions [4.265467042008983]
This paper introduces a new class of loss functions for multi-label classification.
It overcomes disadvantages of commonly used losses such as Hamming and subset 0/1.
The assessment of multi-labels in terms of these losses is illustrated in an empirical study.
arXiv Detail & Related papers (2020-11-02T07:42:15Z) - Pointwise Binary Classification with Pairwise Confidence Comparisons [97.79518780631457]
We propose pairwise comparison (Pcomp) classification, where we have only pairs of unlabeled data that we know one is more likely to be positive than the other.
We link Pcomp classification to noisy-label learning to develop a progressive URE and improve it by imposing consistency regularization.
arXiv Detail & Related papers (2020-10-05T09:23:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.