RankCSE: Unsupervised Sentence Representations Learning via Learning to
Rank
- URL: http://arxiv.org/abs/2305.16726v1
- Date: Fri, 26 May 2023 08:27:07 GMT
- Title: RankCSE: Unsupervised Sentence Representations Learning via Learning to
Rank
- Authors: Jiduan Liu, Jiahao Liu, Qifan Wang, Jingang Wang, Wei Wu, Yunsen Xian,
Dongyan Zhao, Kai Chen, Rui Yan
- Abstract summary: We propose a novel approach, RankCSE, for unsupervised sentence representation learning.
It incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.
An extensive set of experiments are conducted on both semantic textual similarity (STS) and transfer (TR) tasks.
- Score: 54.854714257687334
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised sentence representation learning is one of the fundamental
problems in natural language processing with various downstream applications.
Recently, contrastive learning has been widely adopted which derives
high-quality sentence representations by pulling similar semantics closer and
pushing dissimilar ones away. However, these methods fail to capture the
fine-grained ranking information among the sentences, where each sentence is
only treated as either positive or negative. In many real-world scenarios, one
needs to distinguish and rank the sentences based on their similarities to a
query sentence, e.g., very relevant, moderate relevant, less relevant,
irrelevant, etc. In this paper, we propose a novel approach, RankCSE, for
unsupervised sentence representation learning, which incorporates ranking
consistency and ranking distillation with contrastive learning into a unified
framework. In particular, we learn semantically discriminative sentence
representations by simultaneously ensuring ranking consistency between two
representations with different dropout masks, and distilling listwise ranking
knowledge from the teacher. An extensive set of experiments are conducted on
both semantic textual similarity (STS) and transfer (TR) tasks. Experimental
results demonstrate the superior performance of our approach over several
state-of-the-art baselines.
Related papers
- DenoSent: A Denoising Objective for Self-Supervised Sentence
Representation Learning [59.4644086610381]
We propose a novel denoising objective that inherits from another perspective, i.e., the intra-sentence perspective.
By introducing both discrete and continuous noise, we generate noisy sentences and then train our model to restore them to their original form.
Our empirical evaluations demonstrate that this approach delivers competitive results on both semantic textual similarity (STS) and a wide range of transfer tasks.
arXiv Detail & Related papers (2024-01-24T17:48:45Z) - DebCSE: Rethinking Unsupervised Contrastive Sentence Embedding Learning
in the Debiasing Perspective [1.351603931922027]
We argue that effectively eliminating the influence of various biases is crucial for learning high-quality sentence embeddings.
We propose a novel contrastive framework for sentence embedding, termed DebCSE, which can eliminate the impact of these biases.
arXiv Detail & Related papers (2023-09-14T02:43:34Z) - Towards Unsupervised Recognition of Token-level Semantic Differences in
Related Documents [61.63208012250885]
We formulate recognizing semantic differences as a token-level regression task.
We study three unsupervised approaches that rely on a masked language model.
Our results show that an approach based on word alignment and sentence-level contrastive learning has a robust correlation to gold labels.
arXiv Detail & Related papers (2023-05-22T17:58:04Z) - Conditional Supervised Contrastive Learning for Fair Text Classification [59.813422435604025]
We study learning fair representations that satisfy a notion of fairness known as equalized odds for text classification via contrastive learning.
Specifically, we first theoretically analyze the connections between learning representations with a fairness constraint and conditional supervised contrastive objectives.
arXiv Detail & Related papers (2022-05-23T17:38:30Z) - DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings [51.274478128525686]
DiffCSE is an unsupervised contrastive learning framework for learning sentence embeddings.
Our experiments show that DiffCSE achieves state-of-the-art results among unsupervised sentence representation learning methods.
arXiv Detail & Related papers (2022-04-21T17:32:01Z) - Generative or Contrastive? Phrase Reconstruction for Better Sentence
Representation Learning [86.01683892956144]
We propose a novel generative self-supervised learning objective based on phrase reconstruction.
Our generative learning may yield powerful enough sentence representation and achieve performance in Sentence Textual Similarity tasks on par with contrastive learning.
arXiv Detail & Related papers (2022-04-20T10:00:46Z) - Pair-Level Supervised Contrastive Learning for Natural Language
Inference [11.858875850000457]
We propose a Pair-level Supervised Contrastive Learning approach (PairSCL)
A contrastive learning objective is designed to distinguish the varied classes of sentence pairs by pulling those in one class together and pushing apart the pairs in other classes.
We evaluate PairSCL on two public datasets of NLI where the accuracy of PairSCL outperforms other methods by 2.1% on average.
arXiv Detail & Related papers (2022-01-26T13:34:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.