A Mutually Reinforced Framework for Pretrained Sentence Embeddings
- URL: http://arxiv.org/abs/2202.13802v1
- Date: Mon, 28 Feb 2022 14:00:16 GMT
- Title: A Mutually Reinforced Framework for Pretrained Sentence Embeddings
- Authors: Junhan Yang, Zheng Liu, Shitao Xiao, Jianxun Lian, Lijun Wu, Defu
Lian, Guangzhong Sun, Xing Xie
- Abstract summary: InfoCSE is a novel framework for learning high-quality sentence embeddings.
It exploits the sentence representation model itself and realizes the following iterative self-supervision process.
In other words, the representation learning and data annotation become mutually reinforced, where a strong self-supervision effect can be derived.
- Score: 49.297766436632685
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The lack of labeled data is a major obstacle to learning high-quality
sentence embeddings. Recently, self-supervised contrastive learning (SCL) is
regarded as a promising way to address this problem. However, the existing
works mainly rely on hand-crafted data annotation heuristics to generate
positive training samples, which not only call for domain expertise and
laborious tuning, but are also prone to the following unfavorable cases: 1)
trivial positives, 2) coarse-grained positives, and 3) false positives. As a
result, the self-supervision's quality can be severely limited in reality. In
this work, we propose a novel framework InfoCSE to address the above problems.
Instead of relying on annotation heuristics defined by humans, it leverages the
sentence representation model itself and realizes the following iterative
self-supervision process: on one hand, the improvement of sentence
representation may contribute to the quality of data annotation; on the other
hand, more effective data annotation helps to generate high-quality positive
samples, which will further improve the current sentence representation model.
In other words, the representation learning and data annotation become mutually
reinforced, where a strong self-supervision effect can be derived. Extensive
experiments are performed based on three benchmark datasets, where notable
improvements can be achieved against the existing SCL-based methods.
Related papers
- The Bad Batches: Enhancing Self-Supervised Learning in Image Classification Through Representative Batch Curation [1.519321208145928]
The pursuit of learning robust representations without human supervision is a longstanding challenge.
This paper attempts to alleviate the influence of false positive and false negative pairs by employing pairwise similarity calculations through the Fr'echet ResNet Distance (FRD)
The effectiveness of the proposed method is substantiated by empirical results, where a linear classifier trained on self-supervised contrastive representations achieved an impressive 87.74% top-1 accuracy on STL10 and 99.31% on the Flower102 dataset.
arXiv Detail & Related papers (2024-03-28T17:04:07Z) - Progressive Feature Adjustment for Semi-supervised Learning from
Pretrained Models [39.42802115580677]
Semi-supervised learning (SSL) can leverage both labeled and unlabeled data to build a predictive model.
Recent literature suggests that naively applying state-of-the-art SSL with a pretrained model fails to unleash the full potential of training data.
We propose to use pseudo-labels from the unlabelled data to update the feature extractor that is less sensitive to incorrect labels.
arXiv Detail & Related papers (2023-09-09T01:57:14Z) - Entailment as Robust Self-Learner [14.86757876218415]
We design a prompting strategy that formulates a number of different NLU tasks as contextual entailment.
We propose the Simple Pseudo-Label Editing (SimPLE) algorithm for better pseudo-labeling quality in self-training.
arXiv Detail & Related papers (2023-05-26T18:41:23Z) - Generate, Discriminate and Contrast: A Semi-Supervised Sentence
Representation Learning Framework [68.04940365847543]
We propose a semi-supervised sentence embedding framework, GenSE, that effectively leverages large-scale unlabeled data.
Our method include three parts: 1) Generate: A generator/discriminator model is jointly trained to synthesize sentence pairs from open-domain unlabeled corpus; 2) Discriminate: Noisy sentence pairs are filtered out by the discriminator to acquire high-quality positive and negative sentence pairs; 3) Contrast: A prompt-based contrastive approach is presented for sentence representation learning with both annotated and synthesized data.
arXiv Detail & Related papers (2022-10-30T10:15:21Z) - Contrastive Learning for Fair Representations [50.95604482330149]
Trained classification models can unintentionally lead to biased representations and predictions.
Existing debiasing methods for classification models, such as adversarial training, are often expensive to train and difficult to optimise.
We propose a method for mitigating bias by incorporating contrastive learning, in which instances sharing the same class label are encouraged to have similar representations.
arXiv Detail & Related papers (2021-09-22T10:47:51Z) - Avoiding Inference Heuristics in Few-shot Prompt-based Finetuning [57.4036085386653]
We show that prompt-based models for sentence pair classification tasks still suffer from a common pitfall of adopting inferences based on lexical overlap.
We then show that adding a regularization that preserves pretraining weights is effective in mitigating this destructive tendency of few-shot finetuning.
arXiv Detail & Related papers (2021-09-09T10:10:29Z) - Generic Semi-Supervised Adversarial Subject Translation for Sensor-Based
Human Activity Recognition [6.2997667081978825]
This paper presents a novel generic and robust approach for semi-supervised domain adaptation in Human Activity Recognition.
It capitalizes on the advantages of the adversarial framework to tackle the shortcomings, by leveraging knowledge from annotated samples exclusively from the source subject and unlabeled ones of the target subject.
The results demonstrate the effectiveness of our proposed algorithms over state-of-the-art methods, which led in up to 13%, 4%, and 13% improvement of our high-level activities recognition metrics for Opportunity, LISSI, and PAMAP2 datasets.
arXiv Detail & Related papers (2020-11-11T12:16:23Z) - Mind the Trade-off: Debiasing NLU Models without Degrading the
In-distribution Performance [70.31427277842239]
We introduce a novel debiasing method called confidence regularization.
It discourages models from exploiting biases while enabling them to receive enough incentive to learn from all the training examples.
We evaluate our method on three NLU tasks and show that, in contrast to its predecessors, it improves the performance on out-of-distribution datasets.
arXiv Detail & Related papers (2020-05-01T11:22:55Z) - On Positive-Unlabeled Classification in GAN [130.43248168149432]
This paper defines a positive and unlabeled classification problem for standard GANs.
It then leads to a novel technique to stabilize the training of the discriminator in GANs.
arXiv Detail & Related papers (2020-02-04T05:59:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.