CIL: Contrastive Instance Learning Framework for Distantly Supervised
Relation Extraction
- URL: http://arxiv.org/abs/2106.10855v1
- Date: Mon, 21 Jun 2021 04:51:59 GMT
- Title: CIL: Contrastive Instance Learning Framework for Distantly Supervised
Relation Extraction
- Authors: Tao Chen, Haizhou Shi, Siliang Tang, Zhigang Chen, Fei Wu, Yueting
Zhuang
- Abstract summary: We go beyond typical multi-instance learning (MIL) framework and propose a novel contrastive instance learning (CIL) framework.
Specifically, we regard the initial MIL as the relational triple encoder and constraint positive pairs against negative pairs for each instance.
Experiments demonstrate the effectiveness of our proposed framework, with significant improvements over the previous methods on NYT10, GDS and KBP.
- Score: 52.94486705393062
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The journey of reducing noise from distant supervision (DS) generated
training data has been started since the DS was first introduced into the
relation extraction (RE) task. For the past decade, researchers apply the
multi-instance learning (MIL) framework to find the most reliable feature from
a bag of sentences. Although the pattern of MIL bags can greatly reduce DS
noise, it fails to represent many other useful sentence features in the
datasets. In many cases, these sentence features can only be acquired by extra
sentence-level human annotation with heavy costs. Therefore, the performance of
distantly supervised RE models is bounded. In this paper, we go beyond typical
MIL framework and propose a novel contrastive instance learning (CIL)
framework. Specifically, we regard the initial MIL as the relational triple
encoder and constraint positive pairs against negative pairs for each instance.
Experiments demonstrate the effectiveness of our proposed framework, with
significant improvements over the previous methods on NYT10, GDS and KBP.
Related papers
- Contrastive Learning Is Not Optimal for Quasiperiodic Time Series [4.2807943283312095]
We introduce Distilled Embedding for Almost-Periodic Time Series (DEAPS) in this paper.
DEAPS is a non-contrastive method tailored for quasiperiodic time series, such as electrocardiogram (ECG) data.
We demonstrate a notable improvement of +10% over existing SOTA methods when just a few annotated records are presented to fit a Machine Learning (ML) model.
arXiv Detail & Related papers (2024-07-24T08:02:41Z) - Disperse-Then-Merge: Pushing the Limits of Instruction Tuning via Alignment Tax Reduction [75.25114727856861]
Large language models (LLMs) tend to suffer from deterioration at the latter stage ofSupervised fine-tuning process.
We introduce a simple disperse-then-merge framework to address the issue.
Our framework outperforms various sophisticated methods such as data curation and training regularization on a series of standard knowledge and reasoning benchmarks.
arXiv Detail & Related papers (2024-05-22T08:18:19Z) - Contrastive Multiple Instance Learning for Weakly Supervised Person ReID [50.04900262181093]
We introduce Contrastive Multiple Instance Learning (CMIL), a novel framework tailored for more effective weakly supervised ReID.
CMIL distinguishes itself by requiring only a single model and no pseudo labels while leveraging contrastive losses.
We release the WL-MUDD dataset, an extension of the MUDD dataset featuring naturally occurring weak labels from the real-world application at PerformancePhoto.co.
arXiv Detail & Related papers (2024-02-12T14:48:31Z) - Multiple Instance Learning Framework with Masked Hard Instance Mining
for Whole Slide Image Classification [11.996318969699296]
Masked hard instance mining (MHIM-MIL) is presented.
MHIM-MIL uses a Siamese structure (Teacher-Student) with a consistency constraint to explore potential hard instances.
Experimental results on the CAMELYON-16 and TCGA Lung Cancer datasets demonstrate that MHIM-MIL outperforms other latest methods in terms of performance and training cost.
arXiv Detail & Related papers (2023-07-28T01:40:04Z) - Alleviating Over-smoothing for Unsupervised Sentence Representation [96.19497378628594]
We present a Simple method named Self-Contrastive Learning (SSCL) to alleviate this issue.
Our proposed method is quite simple and can be easily extended to various state-of-the-art models for performance boosting.
arXiv Detail & Related papers (2023-05-09T11:00:02Z) - Latent Feature Relation Consistency for Adversarial Robustness [80.24334635105829]
misclassification will occur when deep neural networks predict adversarial examples which add human-imperceptible adversarial noise to natural examples.
We propose textbfLatent textbfFeature textbfRelation textbfConsistency (textbfLFRC)
LFRC constrains the relation of adversarial examples in latent space to be consistent with the natural examples.
arXiv Detail & Related papers (2023-03-29T13:50:01Z) - ProtoMIL: Multiple Instance Learning with Prototypical Parts for
Fine-Grained Interpretability [2.094672430475796]
Multiple Instance Learning (MIL) gains popularity in many real-life machine learning applications due to its weakly supervised nature.
In this paper, we introduce ProtoMIL, a novel self-explainable MIL method inspired by the case-based reasoning process that operates on visual prototypes.
Thanks to incorporating prototypical features into objects description, ProtoMIL unprecedentedly joins the model accuracy and fine-grained interpretability.
arXiv Detail & Related papers (2021-08-24T10:02:31Z) - A Sample-Based Training Method for Distantly Supervised Relation
Extraction with Pre-Trained Transformers [4.726777092009553]
We propose a novel sampling method for DSRE that relaxes hardware requirements.
In the proposed method, we limit the number of sentences in a batch by randomly sampling sentences from the bags in the batch.
To alleviate the issues caused by random sampling, we use an ensemble of trained models for prediction.
arXiv Detail & Related papers (2021-04-15T15:09:34Z) - Adversarial Examples for Unsupervised Machine Learning Models [71.81480647638529]
Adrial examples causing evasive predictions are widely used to evaluate and improve the robustness of machine learning models.
We propose a framework of generating adversarial examples for unsupervised models and demonstrate novel applications to data augmentation.
arXiv Detail & Related papers (2021-03-02T17:47:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.