Towards Integration of Discriminability and Robustness for
Document-Level Relation Extraction
- URL: http://arxiv.org/abs/2304.00824v1
- Date: Mon, 3 Apr 2023 09:11:18 GMT
- Title: Towards Integration of Discriminability and Robustness for
Document-Level Relation Extraction
- Authors: Jia Guo, Stanley Kok, Lidong Bing
- Abstract summary: Document-level relation extraction (DocRE) predicts relations for entity pairs that rely on long-range context-dependent reasoning in a document.
In this work, we aim to achieve better integration of both the discriminability and robustness for the DocRE problem.
We innovatively customize entropy minimization and supervised contrastive learning for the challenging multi-label and long-tailed learning problems.
- Score: 41.51148745387936
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Document-level relation extraction (DocRE) predicts relations for entity
pairs that rely on long-range context-dependent reasoning in a document. As a
typical multi-label classification problem, DocRE faces the challenge of
effectively distinguishing a small set of positive relations from the majority
of negative ones. This challenge becomes even more difficult to overcome when
there exists a significant number of annotation errors in the dataset. In this
work, we aim to achieve better integration of both the discriminability and
robustness for the DocRE problem. Specifically, we first design an effective
loss function to endow high discriminability to both probabilistic outputs and
internal representations. We innovatively customize entropy minimization and
supervised contrastive learning for the challenging multi-label and long-tailed
learning problems. To ameliorate the impact of label errors, we equipped our
method with a novel negative label sampling strategy to strengthen the model
robustness. In addition, we introduce two new data regimes to mimic more
realistic scenarios with annotation errors and evaluate our sampling strategy.
Experimental results verify the effectiveness of each component and show that
our method achieves new state-of-the-art results on the DocRED dataset, its
recently cleaned version, Re-DocRED, and the proposed data regimes.
Related papers
- Tackling Diverse Minorities in Imbalanced Classification [80.78227787608714]
Imbalanced datasets are commonly observed in various real-world applications, presenting significant challenges in training classifiers.
We propose generating synthetic samples iteratively by mixing data samples from both minority and majority classes.
We demonstrate the effectiveness of our proposed framework through extensive experiments conducted on seven publicly available benchmark datasets.
arXiv Detail & Related papers (2023-08-28T18:48:34Z) - Contrastive Error Attribution for Finetuned Language Models [35.80256755393739]
noisy and misannotated data is a core cause of hallucinations and unfaithful outputs in Natural Language Generation (NLG) tasks.
We introduce a framework to identify and remove low-quality training instances that lead to undesirable outputs.
We show that existing approaches for error tracing, such as gradient-based influence measures, do not perform reliably for detecting faithfulness errors.
arXiv Detail & Related papers (2022-12-21T02:28:07Z) - Temporal Output Discrepancy for Loss Estimation-based Active Learning [65.93767110342502]
We present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss.
Our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.
arXiv Detail & Related papers (2022-12-20T19:29:37Z) - HCL-TAT: A Hybrid Contrastive Learning Method for Few-shot Event
Detection with Task-Adaptive Threshold [18.165302114575212]
We propose a novel Hybrid Contrastive Learning method with a Task-Adaptive Threshold (abbreviated as HCLTAT)
In this paper, we propose a novel Hybrid Contrastive Learning method with a Task-Adaptive Threshold (abbreviated as HCLTAT), which enables discriminative representation learning with a two-view contrastive loss.
Experiments on the benchmark dataset FewEvent demonstrate the superiority of our method to achieve better results compared to the state-of-the-arts.
arXiv Detail & Related papers (2022-10-17T07:37:38Z) - On Modality Bias Recognition and Reduction [70.69194431713825]
We study the modality bias problem in the context of multi-modal classification.
We propose a plug-and-play loss function method, whereby the feature space for each label is adaptively learned.
Our method yields remarkable performance improvements compared with the baselines.
arXiv Detail & Related papers (2022-02-25T13:47:09Z) - Adaptive Affinity Loss and Erroneous Pseudo-Label Refinement for Weakly
Supervised Semantic Segmentation [48.294903659573585]
In this paper, we propose to embed affinity learning of multi-stage approaches in a single-stage model.
A deep neural network is used to deliver comprehensive semantic information in the training phase.
Experiments are conducted on the PASCAL VOC 2012 dataset to evaluate the effectiveness of our proposed approach.
arXiv Detail & Related papers (2021-08-03T07:48:33Z) - Semi-Supervised Active Learning with Temporal Output Discrepancy [42.01906895756629]
We present a novel deep active learning approach that queries the oracle for data annotation when the unlabeled sample is believed to incorporate high loss.
Our approach achieves superior performances than the state-of-the-art active learning methods on image classification and semantic segmentation tasks.
arXiv Detail & Related papers (2021-07-29T16:25:56Z) - WSSOD: A New Pipeline for Weakly- and Semi-Supervised Object Detection [75.80075054706079]
We propose a weakly- and semi-supervised object detection framework (WSSOD)
An agent detector is first trained on a joint dataset and then used to predict pseudo bounding boxes on weakly-annotated images.
The proposed framework demonstrates remarkable performance on PASCAL-VOC and MSCOCO benchmark, achieving a high performance comparable to those obtained in fully-supervised settings.
arXiv Detail & Related papers (2021-05-21T11:58:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.