Toward Interpretable Semantic Textual Similarity via Optimal
Transport-based Contrastive Sentence Learning
- URL: http://arxiv.org/abs/2202.13196v1
- Date: Sat, 26 Feb 2022 17:28:02 GMT
- Title: Toward Interpretable Semantic Textual Similarity via Optimal
Transport-based Contrastive Sentence Learning
- Authors: Seonghyeon Lee, Dongha Lee, Seongbo Jang, Hwanjo Yu
- Abstract summary: We describe the sentence distance as the weighted sum of contextualized token distances on the basis of a transportation problem.
We then present the optimal transport-based distance measure, named RCMD; it identifies and leverages semantically-aligned token pairs.
In the end, we propose CLRCMD, a contrastive learning framework that optimize RCMD of sentence pairs.
- Score: 29.462788855992617
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Recently, finetuning a pretrained language model to capture the similarity
between sentence embeddings has shown the state-of-the-art performance on the
semantic textual similarity (STS) task. However, the absence of an
interpretation method for the sentence similarity makes it difficult to explain
the model output. In this work, we explicitly describe the sentence distance as
the weighted sum of contextualized token distances on the basis of a
transportation problem, and then present the optimal transport-based distance
measure, named RCMD; it identifies and leverages semantically-aligned token
pairs. In the end, we propose CLRCMD, a contrastive learning framework that
optimizes RCMD of sentence pairs, which enhances the quality of sentence
similarity and their interpretation. Extensive experiments demonstrate that our
learning framework outperforms other baselines on both STS and
interpretable-STS benchmarks, indicating that it computes effective sentence
similarity and also provides interpretation consistent with human judgement.
Related papers
- DenoSent: A Denoising Objective for Self-Supervised Sentence
Representation Learning [59.4644086610381]
We propose a novel denoising objective that inherits from another perspective, i.e., the intra-sentence perspective.
By introducing both discrete and continuous noise, we generate noisy sentences and then train our model to restore them to their original form.
Our empirical evaluations demonstrate that this approach delivers competitive results on both semantic textual similarity (STS) and a wide range of transfer tasks.
arXiv Detail & Related papers (2024-01-24T17:48:45Z) - Predicting Text Preference Via Structured Comparative Reasoning [110.49560164568791]
We introduce SC, a prompting approach that predicts text preferences by generating structured intermediate comparisons.
We select consistent comparisons with a pairwise consistency comparator that ensures each aspect's comparisons clearly distinguish differences between texts.
Our comprehensive evaluations across various NLP tasks, including summarization, retrieval, and automatic rating, demonstrate that SC equips LLMs to achieve state-of-the-art performance in text preference prediction.
arXiv Detail & Related papers (2023-11-14T18:51:38Z) - Bridging Continuous and Discrete Spaces: Interpretable Sentence
Representation Learning via Compositional Operations [80.45474362071236]
It is unclear whether the compositional semantics of sentences can be directly reflected as compositional operations in the embedding space.
We propose InterSent, an end-to-end framework for learning interpretable sentence embeddings.
arXiv Detail & Related papers (2023-05-24T00:44:49Z) - Sentence Representations via Gaussian Embedding [15.235687410343171]
GaussCSE is a contrastive learning framework for sentence embedding.
It can handle asymmetric relationships between sentences, along with a similarity measure for identifying inclusion relations.
Our experiments show that GaussCSE achieves the same performance as previous methods in natural language inference tasks.
arXiv Detail & Related papers (2023-05-22T12:51:38Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - SBERT studies Meaning Representations: Decomposing Sentence Embeddings
into Explainable AMR Meaning Features [22.8438857884398]
We create similarity metrics that are highly effective, while also providing an interpretable rationale for their rating.
Our approach works in two steps: We first select AMR graph metrics that measure meaning similarity of sentences with respect to key semantic facets.
Second, we employ these metrics to induce Semantically Structured Sentence BERT embeddings, which are composed of different meaning aspects captured in different sub-spaces.
arXiv Detail & Related papers (2022-06-14T17:37:18Z) - Conditional Supervised Contrastive Learning for Fair Text Classification [59.813422435604025]
We study learning fair representations that satisfy a notion of fairness known as equalized odds for text classification via contrastive learning.
Specifically, we first theoretically analyze the connections between learning representations with a fairness constraint and conditional supervised contrastive objectives.
arXiv Detail & Related papers (2022-05-23T17:38:30Z) - Comparison and Combination of Sentence Embeddings Derived from Different
Supervision Signals [20.853681115929422]
We focus on two types of sentence embeddings obtained by using natural language inference (NLI) datasets and definition sentences from a word dictionary.
We compare their performance with the semantic textual similarity (STS) task using the STS data partitioned by two perspectives.
We also demonstrate that combining the two types of embeddings yields substantially better performances than respective models on unsupervised STS tasks and downstream tasks.
arXiv Detail & Related papers (2022-02-07T08:15:48Z) - Contextualized Semantic Distance between Highly Overlapped Texts [85.1541170468617]
Overlapping frequently occurs in paired texts in natural language processing tasks like text editing and semantic similarity evaluation.
This paper aims to address the issue with a mask-and-predict strategy.
We take the words in the longest common sequence as neighboring words and use masked language modeling (MLM) to predict the distributions on their positions.
Experiments on Semantic Textual Similarity show NDD to be more sensitive to various semantic differences, especially on highly overlapped paired texts.
arXiv Detail & Related papers (2021-10-04T03:59:15Z) - Structural-Aware Sentence Similarity with Recursive Optimal Transport [11.052550499042646]
We develop Recursive Optimal Similarity (ROTS) for sentences with the valuable semantic insights from cosine similarity of weighted average of word vectors and optimal transport.
Our experiments over 20 sentence textural similarity (STS) datasets show the clear advantage of ROTS over all weakly supervised approaches.
arXiv Detail & Related papers (2020-01-28T09:07:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.