D2CSE: Difference-aware Deep continuous prompts for Contrastive Sentence
Embeddings
- URL: http://arxiv.org/abs/2304.08991v1
- Date: Tue, 18 Apr 2023 13:45:07 GMT
- Title: D2CSE: Difference-aware Deep continuous prompts for Contrastive Sentence
Embeddings
- Authors: Hyunjae Lee
- Abstract summary: This paper describes Difference-aware Deep continuous prompt for Contrastive Sentence Embeddings (D2CSE) that learns sentence embeddings.
Compared to state-of-the-art approaches, D2CSE computes sentence vectors that are exceptional to distinguish a subtle difference in similar sentences.
- Score: 3.04585143845864
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper describes Difference-aware Deep continuous prompt for Contrastive
Sentence Embeddings (D2CSE) that learns sentence embeddings. Compared to
state-of-the-art approaches, D2CSE computes sentence vectors that are
exceptional to distinguish a subtle difference in similar sentences by
employing a simple neural architecture for continuous prompts. Unlike existing
architectures that require multiple pretrained language models (PLMs) to
process a pair of the original and corrupted (subtly modified) sentences, D2CSE
avoids cumbersome fine-tuning of multiple PLMs by only optimizing continuous
prompts by performing multiple tasks -- i.e., contrastive learning and
conditional replaced token detection all done in a self-guided manner. D2CSE
overloads a single PLM on continuous prompts and greatly saves memory
consumption as a result. The number of training parameters in D2CSE is reduced
to about 1\% of existing approaches while substantially improving the quality
of sentence embeddings. We evaluate D2CSE on seven Semantic Textual Similarity
(STS) benchmarks, using three different metrics, namely, Spearman's rank
correlation, recall@K for a retrieval task, and the anisotropy of an embedding
space measured in alignment and uniformity. Our empirical results suggest that
shallow (not too meticulously devised) continuous prompts can be honed
effectively for multiple NLP tasks and lead to improvements upon existing
state-of-the-art approaches.
Related papers
- Bidirectional Decoding: Improving Action Chunking via Closed-Loop Resampling [51.38330727868982]
Bidirectional Decoding (BID) is a test-time inference algorithm that bridges action chunking with closed-loop operations.
We show that BID boosts the performance of two state-of-the-art generative policies across seven simulation benchmarks and two real-world tasks.
arXiv Detail & Related papers (2024-08-30T15:39:34Z) - Aligning Speakers: Evaluating and Visualizing Text-based Diarization
Using Efficient Multiple Sequence Alignment (Extended Version) [21.325463387256807]
Two new metrics are proposed, Text-based Diarization Error Rate and Diarization F1, which perform utterance- and word-level evaluations.
Our metrics encompass more types of errors compared to existing ones, allowing us to make a more comprehensive analysis in speaker diarization.
arXiv Detail & Related papers (2023-09-14T12:43:26Z) - Deep Declarative Dynamic Time Warping for End-to-End Learning of
Alignment Paths [54.53208538517505]
This paper addresses learning end-to-end models for time series data that include a temporal alignment step via dynamic time warping (DTW)
We propose a DTW layer based around bi-level optimisation and deep declarative networks, which we name DecDTW.
We show that this property is particularly useful for applications where downstream loss functions are defined on the optimal alignment path itself.
arXiv Detail & Related papers (2023-03-19T21:58:37Z) - DiffCSE: Difference-based Contrastive Learning for Sentence Embeddings [51.274478128525686]
DiffCSE is an unsupervised contrastive learning framework for learning sentence embeddings.
Our experiments show that DiffCSE achieves state-of-the-art results among unsupervised sentence representation learning methods.
arXiv Detail & Related papers (2022-04-21T17:32:01Z) - Fine-grained Temporal Contrastive Learning for Weakly-supervised
Temporal Action Localization [87.47977407022492]
This paper argues that learning by contextually comparing sequence-to-sequence distinctions offers an essential inductive bias in weakly-supervised action localization.
Under a differentiable dynamic programming formulation, two complementary contrastive objectives are designed, including Fine-grained Sequence Distance (FSD) contrasting and Longest Common Subsequence (LCS) contrasting.
Our method achieves state-of-the-art performance on two popular benchmarks.
arXiv Detail & Related papers (2022-03-31T05:13:50Z) - Benchmarking Deep Models for Salient Object Detection [67.07247772280212]
We construct a general SALient Object Detection (SALOD) benchmark to conduct a comprehensive comparison among several representative SOD methods.
In the above experiments, we find that existing loss functions usually specialized in some metrics but reported inferior results on the others.
We propose a novel Edge-Aware (EA) loss that promotes deep networks to learn more discriminative features by integrating both pixel- and image-level supervision signals.
arXiv Detail & Related papers (2022-02-07T03:43:16Z) - Sentence Embeddings using Supervised Contrastive Learning [0.0]
We propose a new method to build sentence embeddings by doing supervised contrastive learning.
Our method fine-tunes pretrained BERT on SNLI data, incorporating both supervised crossentropy loss and supervised contrastive loss.
arXiv Detail & Related papers (2021-06-09T03:30:29Z) - Boosting Continuous Sign Language Recognition via Cross Modality
Augmentation [135.30357113518127]
Continuous sign language recognition deals with unaligned video-text pair.
We propose a novel architecture with cross modality augmentation.
The proposed framework can be easily extended to other existing CTC based continuous SLR architectures.
arXiv Detail & Related papers (2020-10-11T15:07:50Z) - ReADS: A Rectified Attentional Double Supervised Network for Scene Text
Recognition [22.367624178280682]
We elaborately design a Rectified Attentional Double Supervised Network (ReADS) for general scene text recognition.
The ReADS can be trained end-to-end and only word-level annotations are required.
arXiv Detail & Related papers (2020-04-05T02:05:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.