SBERT studies Meaning Representations: Decomposing Sentence Embeddings
into Explainable AMR Meaning Features
- URL: http://arxiv.org/abs/2206.07023v1
- Date: Tue, 14 Jun 2022 17:37:18 GMT
- Title: SBERT studies Meaning Representations: Decomposing Sentence Embeddings
into Explainable AMR Meaning Features
- Authors: Juri Opitz and Anette Frank
- Abstract summary: We create similarity metrics that are highly effective, while also providing an interpretable rationale for their rating.
Our approach works in two steps: We first select AMR graph metrics that measure meaning similarity of sentences with respect to key semantic facets.
Second, we employ these metrics to induce Semantically Structured Sentence BERT embeddings, which are composed of different meaning aspects captured in different sub-spaces.
- Score: 22.8438857884398
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Metrics for graph-based meaning representations (e.g., Abstract Meaning
Representation, AMR) can help us uncover key semantic aspects in which two
sentences are similar to each other. However, such metrics tend to be slow,
rely on parsers, and do not reach state-of-the-art performance when rating
sentence similarity. On the other hand, models based on large-pretrained
language models, such as S(entence)BERT, show high correlation to human
similarity ratings, but lack interpretability.
In this paper, we aim at the best of these two worlds, by creating similarity
metrics that are highly effective, while also providing an interpretable
rationale for their rating. Our approach works in two steps: We first select
AMR graph metrics that measure meaning similarity of sentences with respect to
key semantic facets, such as, i.a., semantic roles, negation, or
quantification. Second, we employ these metrics to induce Semantically
Structured Sentence BERT embeddings (S$^3$BERT), which are composed of
different meaning aspects captured in different sub-spaces. In our experimental
studies, we show that our approach offers a valuable balance between
performance and interpretability.
Related papers
- Rematch: Robust and Efficient Matching of Local Knowledge Graphs to Improve Structural and Semantic Similarity [6.1980259703476674]
We introduce a novel AMR similarity metric, rematch, alongside a new evaluation for structural similarity called RARE.
Rematch ranks second in structural similarity; and first in semantic similarity by 1--5 percentage points on the STS-B and SICK-R benchmarks.
arXiv Detail & Related papers (2024-04-02T17:33:00Z) - Cobra Effect in Reference-Free Image Captioning Metrics [58.438648377314436]
A proliferation of reference-free methods, leveraging visual-language pre-trained models (VLMs), has emerged.
In this paper, we study if there are any deficiencies in reference-free metrics.
We employ GPT-4V as an evaluative tool to assess generated sentences and the result reveals that our approach achieves state-of-the-art (SOTA) performance.
arXiv Detail & Related papers (2024-02-18T12:36:23Z) - AMR4NLI: Interpretable and robust NLI measures from semantic graphs [28.017617759762278]
Natural language inference asks whether a given premise entails a given hypothesis.
We compare semantic structures to represent premise and hypothesis, including sets of contextualized embeddings and semantic graphs.
Our evaluation finds value in both contextualized embeddings and semantic graphs.
arXiv Detail & Related papers (2023-06-01T17:39:40Z) - RankCSE: Unsupervised Sentence Representations Learning via Learning to
Rank [54.854714257687334]
We propose a novel approach, RankCSE, for unsupervised sentence representation learning.
It incorporates ranking consistency and ranking distillation with contrastive learning into a unified framework.
An extensive set of experiments are conducted on both semantic textual similarity (STS) and transfer (TR) tasks.
arXiv Detail & Related papers (2023-05-26T08:27:07Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - Retrofitting Multilingual Sentence Embeddings with Abstract Meaning
Representation [70.58243648754507]
We introduce a new method to improve existing multilingual sentence embeddings with Abstract Meaning Representation (AMR)
Compared with the original textual input, AMR is a structured semantic representation that presents the core concepts and relations in a sentence explicitly and unambiguously.
Experiment results show that retrofitting multilingual sentence embeddings with AMR leads to better state-of-the-art performance on both semantic similarity and transfer tasks.
arXiv Detail & Related papers (2022-10-18T11:37:36Z) - Measuring Fine-Grained Semantic Equivalence with Abstract Meaning
Representation [9.666975331506812]
Identifying semantically equivalent sentences is important for many NLP tasks.
Current approaches to semantic equivalence take a loose, sentence-level approach to "equivalence"
We introduce a novel, more sensitive method of characterizing semantic equivalence that leverages Abstract Meaning Representation graph structures.
arXiv Detail & Related papers (2022-10-06T16:08:27Z) - Contextualized Semantic Distance between Highly Overlapped Texts [85.1541170468617]
Overlapping frequently occurs in paired texts in natural language processing tasks like text editing and semantic similarity evaluation.
This paper aims to address the issue with a mask-and-predict strategy.
We take the words in the longest common sequence as neighboring words and use masked language modeling (MLM) to predict the distributions on their positions.
Experiments on Semantic Textual Similarity show NDD to be more sensitive to various semantic differences, especially on highly overlapped paired texts.
arXiv Detail & Related papers (2021-10-04T03:59:15Z) - Making Better Use of Bilingual Information for Cross-Lingual AMR Parsing [88.08581016329398]
We argue that the misprediction of concepts is due to the high relevance between English tokens and AMR concepts.
We introduce bilingual input, namely the translated texts as well as non-English texts, in order to enable the model to predict more accurate concepts.
arXiv Detail & Related papers (2021-06-09T05:14:54Z) - AMR Similarity Metrics from Principles [21.915057426589748]
We establish criteria that enable researchers to perform a principled assessment of metrics comparing meaning representations like AMR.
We propose a novel metric S$2$match that is more benevolent to only very slight meaning deviations and targets the fulfilment of all established criteria.
arXiv Detail & Related papers (2020-01-29T16:19:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.