Structural-Aware Sentence Similarity with Recursive Optimal Transport
- URL: http://arxiv.org/abs/2002.00745v1
- Date: Tue, 28 Jan 2020 09:07:47 GMT
- Title: Structural-Aware Sentence Similarity with Recursive Optimal Transport
- Authors: Zihao Wang, Yong Zhang, Hao Wu
- Abstract summary: We develop Recursive Optimal Similarity (ROTS) for sentences with the valuable semantic insights from cosine similarity of weighted average of word vectors and optimal transport.
Our experiments over 20 sentence textural similarity (STS) datasets show the clear advantage of ROTS over all weakly supervised approaches.
- Score: 11.052550499042646
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Measuring sentence similarity is a classic topic in natural language
processing. Light-weighted similarities are still of particular practical
significance even when deep learning models have succeeded in many other tasks.
Some light-weighted similarities with more theoretical insights have been
demonstrated to be even stronger than supervised deep learning approaches.
However, the successful light-weighted models such as Word Mover's Distance
[Kusner et al., 2015] or Smooth Inverse Frequency [Arora et al., 2017] failed
to detect the difference from the structure of sentences, i.e. order of words.
To address this issue, we present Recursive Optimal Transport (ROT) framework
to incorporate the structural information with the classic OT. Moreover, we
further develop Recursive Optimal Similarity (ROTS) for sentences with the
valuable semantic insights from the connections between cosine similarity of
weighted average of word vectors and optimal transport. ROTS is
structural-aware and with low time complexity compared to optimal transport.
Our experiments over 20 sentence textural similarity (STS) datasets show the
clear advantage of ROTS over all weakly supervised approaches. Detailed
ablation study demonstrate the effectiveness of ROT and the semantic insights.
Related papers
- GSSF: Generalized Structural Sparse Function for Deep Cross-modal Metric Learning [51.677086019209554]
We propose a Generalized Structural Sparse to capture powerful relationships across modalities for pair-wise similarity learning.
The distance metric delicately encapsulates two formats of diagonal and block-diagonal terms.
Experiments on cross-modal and two extra uni-modal retrieval tasks have validated its superiority and flexibility.
arXiv Detail & Related papers (2024-10-20T03:45:50Z) - Predicting Text Preference Via Structured Comparative Reasoning [110.49560164568791]
We introduce SC, a prompting approach that predicts text preferences by generating structured intermediate comparisons.
We select consistent comparisons with a pairwise consistency comparator that ensures each aspect's comparisons clearly distinguish differences between texts.
Our comprehensive evaluations across various NLP tasks, including summarization, retrieval, and automatic rating, demonstrate that SC equips LLMs to achieve state-of-the-art performance in text preference prediction.
arXiv Detail & Related papers (2023-11-14T18:51:38Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - Unsupervised Sentence Textual Similarity with Compositional Phrase
Semantics [8.729329792251578]
Measuring Sentence Textual Similarity (STS) is a classic task that can be applied to many downstream NLP applications.
In this paper, we focus on unsupervised STS that works on various domains but only requires minimal data and computational resources.
arXiv Detail & Related papers (2022-10-05T14:14:04Z) - Unbalanced CO-Optimal Transport [16.9451175221198]
CO-optimal transport (COOT) takes this comparison further by inferring an alignment between features as well.
We show that it is sensitive to outliers that are omnipresent in real-world data.
This prompts us to propose unbalanced COOT for which we provably show its robustness to noise.
arXiv Detail & Related papers (2022-05-30T08:43:19Z) - HiURE: Hierarchical Exemplar Contrastive Learning for Unsupervised
Relation Extraction [60.80849503639896]
Unsupervised relation extraction aims to extract the relationship between entities from natural language sentences without prior information on relational scope or distribution.
We propose a novel contrastive learning framework named HiURE, which has the capability to derive hierarchical signals from relational feature space using cross hierarchy attention.
Experimental results on two public datasets demonstrate the advanced effectiveness and robustness of HiURE on unsupervised relation extraction when compared with state-of-the-art models.
arXiv Detail & Related papers (2022-05-04T17:56:48Z) - Attributable Visual Similarity Learning [90.69718495533144]
This paper proposes an attributable visual similarity learning (AVSL) framework for a more accurate and explainable similarity measure between images.
Motivated by the human semantic similarity cognition, we propose a generalized similarity learning paradigm to represent the similarity between two images with a graph.
Experiments on the CUB-200-2011, Cars196, and Stanford Online Products datasets demonstrate significant improvements over existing deep similarity learning methods.
arXiv Detail & Related papers (2022-03-28T17:35:31Z) - Toward Interpretable Semantic Textual Similarity via Optimal
Transport-based Contrastive Sentence Learning [29.462788855992617]
We describe the sentence distance as the weighted sum of contextualized token distances on the basis of a transportation problem.
We then present the optimal transport-based distance measure, named RCMD; it identifies and leverages semantically-aligned token pairs.
In the end, we propose CLRCMD, a contrastive learning framework that optimize RCMD of sentence pairs.
arXiv Detail & Related papers (2022-02-26T17:28:02Z) - A Comparative Study on Structural and Semantic Properties of Sentence
Embeddings [77.34726150561087]
We propose a set of experiments using a widely-used large-scale data set for relation extraction.
We show that different embedding spaces have different degrees of strength for the structural and semantic properties.
These results provide useful information for developing embedding-based relation extraction methods.
arXiv Detail & Related papers (2020-09-23T15:45:32Z) - Geometric Dataset Distances via Optimal Transport [15.153110906331733]
We propose an alternative notion of distance between datasets that (i) is model-agnostic, (ii) does not involve training, (iii) can compare datasets even if their label sets are completely disjoint and (iv) has solid theoretical footing.
This distance relies on optimal transport, which provides it with rich geometry awareness, interpretable correspondences and well-understood properties.
Our results show that this novel distance provides meaningful comparison of datasets, and correlates well with transfer learning hardness across various experimental settings and datasets.
arXiv Detail & Related papers (2020-02-07T17:51:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.