Comparison and Combination of Sentence Embeddings Derived from Different
Supervision Signals
- URL: http://arxiv.org/abs/2202.02990v1
- Date: Mon, 7 Feb 2022 08:15:48 GMT
- Title: Comparison and Combination of Sentence Embeddings Derived from Different
Supervision Signals
- Authors: Hayato Tsukagoshi, Ryohei Sasano, Koichi Takeda
- Abstract summary: We focus on two types of sentence embeddings obtained by using natural language inference (NLI) datasets and definition sentences from a word dictionary.
We compare their performance with the semantic textual similarity (STS) task using the STS data partitioned by two perspectives.
We also demonstrate that combining the two types of embeddings yields substantially better performances than respective models on unsupervised STS tasks and downstream tasks.
- Score: 20.853681115929422
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We have recently seen many successful applications of sentence embedding
methods. It has not been well understood, however, what kind of properties are
captured in the resulting sentence embeddings, depending on the supervision
signals. In this paper, we focus on two types of sentence embeddings obtained
by using natural language inference (NLI) datasets and definition sentences
from a word dictionary and investigate their properties by comparing their
performance with the semantic textual similarity (STS) task using the STS data
partitioned by two perspectives: 1) the sources of sentences, and 2) the
superficial similarity of the sentence pairs, and their performance on the
downstream and probing tasks. We also demonstrate that combining the two types
of embeddings yields substantially better performances than respective models
on unsupervised STS tasks and downstream tasks.
Related papers
- DenoSent: A Denoising Objective for Self-Supervised Sentence
Representation Learning [59.4644086610381]
We propose a novel denoising objective that inherits from another perspective, i.e., the intra-sentence perspective.
By introducing both discrete and continuous noise, we generate noisy sentences and then train our model to restore them to their original form.
Our empirical evaluations demonstrate that this approach delivers competitive results on both semantic textual similarity (STS) and a wide range of transfer tasks.
arXiv Detail & Related papers (2024-01-24T17:48:45Z) - RankCSE: Unsupervised Sentence Representations Learning via Learning to
Rank [54.854714257687334]
We propose a novel approach, RankCSE, for unsupervised sentence representation learning.
It incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.
An extensive set of experiments are conducted on both semantic textual similarity (STS) and transfer (TR) tasks.
arXiv Detail & Related papers (2023-05-26T08:27:07Z) - Sentence Representations via Gaussian Embedding [15.235687410343171]
GaussCSE is a contrastive learning framework for sentence embedding.
It can handle asymmetric relationships between sentences, along with a similarity measure for identifying inclusion relations.
Our experiments show that GaussCSE achieves the same performance as previous methods in natural language inference tasks.
arXiv Detail & Related papers (2023-05-22T12:51:38Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - InfoCSE: Information-aggregated Contrastive Learning of Sentence
Embeddings [61.77760317554826]
This paper proposes an information-d contrastive learning framework for learning unsupervised sentence embeddings, termed InfoCSE.
We evaluate the proposed InfoCSE on several benchmark datasets w.r.t the semantic text similarity (STS) task.
Experimental results show that InfoCSE outperforms SimCSE by an average Spearman correlation of 2.60% on BERT-base, and 1.77% on BERT-large.
arXiv Detail & Related papers (2022-10-08T15:53:19Z) - SBERT studies Meaning Representations: Decomposing Sentence Embeddings
into Explainable AMR Meaning Features [22.8438857884398]
We create similarity metrics that are highly effective, while also providing an interpretable rationale for their rating.
Our approach works in two steps: We first select AMR graph metrics that measure meaning similarity of sentences with respect to key semantic facets.
Second, we employ these metrics to induce Semantically Structured Sentence BERT embeddings, which are composed of different meaning aspects captured in different sub-spaces.
arXiv Detail & Related papers (2022-06-14T17:37:18Z) - Toward Interpretable Semantic Textual Similarity via Optimal
Transport-based Contrastive Sentence Learning [29.462788855992617]
We describe the sentence distance as the weighted sum of contextualized token distances on the basis of a transportation problem.
We then present the optimal transport-based distance measure, named RCMD; it identifies and leverages semantically-aligned token pairs.
In the end, we propose CLRCMD, a contrastive learning framework that optimize RCMD of sentence pairs.
arXiv Detail & Related papers (2022-02-26T17:28:02Z) - Pair-Level Supervised Contrastive Learning for Natural Language
Inference [11.858875850000457]
We propose a Pair-level Supervised Contrastive Learning approach (PairSCL)
A contrastive learning objective is designed to distinguish the varied classes of sentence pairs by pulling those in one class together and pushing apart the pairs in other classes.
We evaluate PairSCL on two public datasets of NLI where the accuracy of PairSCL outperforms other methods by 2.1% on average.
arXiv Detail & Related papers (2022-01-26T13:34:52Z) - A Comparative Study on Structural and Semantic Properties of Sentence
Embeddings [77.34726150561087]
We propose a set of experiments using a widely-used large-scale data set for relation extraction.
We show that different embedding spaces have different degrees of strength for the structural and semantic properties.
These results provide useful information for developing embedding-based relation extraction methods.
arXiv Detail & Related papers (2020-09-23T15:45:32Z) - BURT: BERT-inspired Universal Representation from Twin Structure [89.82415322763475]
BURT (BERT inspired Universal Representation from Twin Structure) is capable of generating universal, fixed-size representations for input sequences of any granularity.
Our proposed BURT adopts the Siamese network, learning sentence-level representations from natural language inference dataset and word/phrase-level representations from paraphrasing dataset.
We evaluate BURT across different granularities of text similarity tasks, including STS tasks, SemEval2013 Task 5(a) and some commonly used word similarity tasks.
arXiv Detail & Related papers (2020-04-29T04:01:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.