Generate, Discriminate and Contrast: A Semi-Supervised Sentence
Representation Learning Framework
- URL: http://arxiv.org/abs/2210.16798v1
- Date: Sun, 30 Oct 2022 10:15:21 GMT
- Title: Generate, Discriminate and Contrast: A Semi-Supervised Sentence
Representation Learning Framework
- Authors: Yiming Chen, Yan Zhang, Bin Wang, Zuozhu Liu, Haizhou Li
- Abstract summary: We propose a semi-supervised sentence embedding framework, GenSE, that effectively leverages large-scale unlabeled data.
Our method include three parts: 1) Generate: A generator/discriminator model is jointly trained to synthesize sentence pairs from open-domain unlabeled corpus; 2) Discriminate: Noisy sentence pairs are filtered out by the discriminator to acquire high-quality positive and negative sentence pairs; 3) Contrast: A prompt-based contrastive approach is presented for sentence representation learning with both annotated and synthesized data.
- Score: 68.04940365847543
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most sentence embedding techniques heavily rely on expensive human-annotated
sentence pairs as the supervised signals. Despite the use of large-scale
unlabeled data, the performance of unsupervised methods typically lags far
behind that of the supervised counterparts in most downstream tasks. In this
work, we propose a semi-supervised sentence embedding framework, GenSE, that
effectively leverages large-scale unlabeled data. Our method include three
parts: 1) Generate: A generator/discriminator model is jointly trained to
synthesize sentence pairs from open-domain unlabeled corpus; 2) Discriminate:
Noisy sentence pairs are filtered out by the discriminator to acquire
high-quality positive and negative sentence pairs; 3) Contrast: A prompt-based
contrastive approach is presented for sentence representation learning with
both annotated and synthesized data. Comprehensive experiments show that GenSE
achieves an average correlation score of 85.19 on the STS datasets and
consistent performance improvement on four domain adaptation tasks,
significantly surpassing the state-of-the-art methods and convincingly
corroborating its effectiveness and generalization ability.Code, Synthetic data
and Models available at https://github.com/MatthewCYM/GenSE.
Related papers
- Prototypical Contrastive Learning through Alignment and Uniformity for
Recommendation [6.790779112538357]
We present underlinePrototypical contrastive learning through underlineAlignment and underlineUniformity for recommendation.
Specifically, we first propose prototypes as a latent space to ensure consistency across different augmentations from the origin graph.
The absence of explicit negatives means that directly optimizing the consistency loss between instance and prototype could easily result in dimensional collapse issues.
arXiv Detail & Related papers (2024-02-03T08:19:26Z) - DenoSent: A Denoising Objective for Self-Supervised Sentence
Representation Learning [59.4644086610381]
We propose a novel denoising objective that inherits from another perspective, i.e., the intra-sentence perspective.
By introducing both discrete and continuous noise, we generate noisy sentences and then train our model to restore them to their original form.
Our empirical evaluations demonstrate that this approach delivers competitive results on both semantic textual similarity (STS) and a wide range of transfer tasks.
arXiv Detail & Related papers (2024-01-24T17:48:45Z) - SenTest: Evaluating Robustness of Sentence Encoders [0.4194295877935868]
This work focuses on evaluating the robustness of the sentence encoders.
We employ several adversarial attacks to evaluate its robustness.
The results of the experiments strongly undermine the robustness of sentence encoders.
arXiv Detail & Related papers (2023-11-29T15:21:35Z) - Noisy-Correspondence Learning for Text-to-Image Person Re-identification [50.07634676709067]
We propose a novel Robust Dual Embedding method (RDE) to learn robust visual-semantic associations even with noisy correspondences.
Our method achieves state-of-the-art results both with and without synthetic noisy correspondences on three datasets.
arXiv Detail & Related papers (2023-08-19T05:34:13Z) - SDA: Simple Discrete Augmentation for Contrastive Sentence Representation Learning [14.028140579482688]
SimCSE surprisingly dominates discrete augmentations such as cropping, word deletion, and synonym replacement as reported.
We develop three simple yet effective discrete sentence augmentation schemes: punctuation insertion, modal verbs, and double negation.
Results support the superiority of the proposed methods consistently.
arXiv Detail & Related papers (2022-10-08T08:07:47Z) - Non-contrastive representation learning for intervals from well logs [58.70164460091879]
The representation learning problem in the oil & gas industry aims to construct a model that provides a representation based on logging data for a well interval.
One of the possible approaches is self-supervised learning (SSL)
We are the first to introduce non-contrastive SSL for well-logging data.
arXiv Detail & Related papers (2022-09-28T13:27:10Z) - Adversarial Dual-Student with Differentiable Spatial Warping for
Semi-Supervised Semantic Segmentation [70.2166826794421]
We propose a differentiable geometric warping to conduct unsupervised data augmentation.
We also propose a novel adversarial dual-student framework to improve the Mean-Teacher.
Our solution significantly improves the performance and state-of-the-art results are achieved on both datasets.
arXiv Detail & Related papers (2022-03-05T17:36:17Z) - A Mutually Reinforced Framework for Pretrained Sentence Embeddings [49.297766436632685]
InfoCSE is a novel framework for learning high-quality sentence embeddings.
It exploits the sentence representation model itself and realizes the following iterative self-supervision process.
In other words, the representation learning and data annotation become mutually reinforced, where a strong self-supervision effect can be derived.
arXiv Detail & Related papers (2022-02-28T14:00:16Z) - SentPWNet: A Unified Sentence Pair Weighting Network for Task-specific
Sentence Embedding [12.020634125787279]
We propose a unified locality weighting and learning framework to learn task-specific sentence embedding.
Our model, SentPWNet, exploits the neighboring spatial distribution of each sentence as locality weight to indicate the informative level of sentence pair.
arXiv Detail & Related papers (2020-05-22T18:32:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.