Sentence Representations via Gaussian Embedding
- URL: http://arxiv.org/abs/2305.12990v2
- Date: Tue, 20 Feb 2024 14:23:58 GMT
- Title: Sentence Representations via Gaussian Embedding
- Authors: Shohei Yoda, Hayato Tsukagoshi, Ryohei Sasano, Koichi Takeda
- Abstract summary: GaussCSE is a contrastive learning framework for sentence embedding.
It can handle asymmetric relationships between sentences, along with a similarity measure for identifying inclusion relations.
Our experiments show that GaussCSE achieves the same performance as previous methods in natural language inference tasks.
- Score: 15.235687410343171
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent progress in sentence embedding, which represents the meaning of a
sentence as a point in a vector space, has achieved high performance on tasks
such as a semantic textual similarity (STS) task. However, sentence
representations as a point in a vector space can express only a part of the
diverse information that sentences have, such as asymmetrical relationships
between sentences. This paper proposes GaussCSE, a Gaussian distribution-based
contrastive learning framework for sentence embedding that can handle
asymmetric relationships between sentences, along with a similarity measure for
identifying inclusion relations. Our experiments show that GaussCSE achieves
the same performance as previous methods in natural language inference tasks,
and is able to estimate the direction of entailment relations, which is
difficult with point representations.
Related papers
- DenoSent: A Denoising Objective for Self-Supervised Sentence
Representation Learning [59.4644086610381]
We propose a novel denoising objective that inherits from another perspective, i.e., the intra-sentence perspective.
By introducing both discrete and continuous noise, we generate noisy sentences and then train our model to restore them to their original form.
Our empirical evaluations demonstrate that this approach delivers competitive results on both semantic textual similarity (STS) and a wide range of transfer tasks.
arXiv Detail & Related papers (2024-01-24T17:48:45Z) - RankCSE: Unsupervised Sentence Representations Learning via Learning to
Rank [54.854714257687334]
We propose a novel approach, RankCSE, for unsupervised sentence representation learning.
It incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.
An extensive set of experiments are conducted on both semantic textual similarity (STS) and transfer (TR) tasks.
arXiv Detail & Related papers (2023-05-26T08:27:07Z) - Bridging Continuous and Discrete Spaces: Interpretable Sentence
Representation Learning via Compositional Operations [80.45474362071236]
It is unclear whether the compositional semantics of sentences can be directly reflected as compositional operations in the embedding space.
We propose InterSent, an end-to-end framework for learning interpretable sentence embeddings.
arXiv Detail & Related papers (2023-05-24T00:44:49Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - Subspace Representations for Soft Set Operations and Sentence Similarities [17.52824249186434]
We realize the representation of word sets and corresponding set operations within pre-trained word embedding spaces.
By grounding our approach in the linear subspaces, we enable efficient computation of various set operations.
We show that our subspace-based set operations consistently outperform vector-based ones in both sentence similarity and set retrieval tasks.
arXiv Detail & Related papers (2022-10-24T08:34:10Z) - Toward Interpretable Semantic Textual Similarity via Optimal
Transport-based Contrastive Sentence Learning [29.462788855992617]
We describe the sentence distance as the weighted sum of contextualized token distances on the basis of a transportation problem.
We then present the optimal transport-based distance measure, named RCMD; it identifies and leverages semantically-aligned token pairs.
In the end, we propose CLRCMD, a contrastive learning framework that optimize RCMD of sentence pairs.
arXiv Detail & Related papers (2022-02-26T17:28:02Z) - Comparison and Combination of Sentence Embeddings Derived from Different
Supervision Signals [20.853681115929422]
We focus on two types of sentence embeddings obtained by using natural language inference (NLI) datasets and definition sentences from a word dictionary.
We compare their performance with the semantic textual similarity (STS) task using the STS data partitioned by two perspectives.
We also demonstrate that combining the two types of embeddings yields substantially better performances than respective models on unsupervised STS tasks and downstream tasks.
arXiv Detail & Related papers (2022-02-07T08:15:48Z) - On the Sentence Embeddings from Pre-trained Language Models [78.45172445684126]
In this paper, we argue that the semantic information in the BERT embeddings is not fully exploited.
We find that BERT always induces a non-smooth anisotropic semantic space of sentences, which harms its performance of semantic similarity.
We propose to transform the anisotropic sentence embedding distribution to a smooth and isotropic Gaussian distribution through normalizing flows that are learned with an unsupervised objective.
arXiv Detail & Related papers (2020-11-02T13:14:57Z) - A Comparative Study on Structural and Semantic Properties of Sentence
Embeddings [77.34726150561087]
We propose a set of experiments using a widely-used large-scale data set for relation extraction.
We show that different embedding spaces have different degrees of strength for the structural and semantic properties.
These results provide useful information for developing embedding-based relation extraction methods.
arXiv Detail & Related papers (2020-09-23T15:45:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.