Sentence Similarity Based on Contexts
- URL: http://arxiv.org/abs/2105.07623v1
- Date: Mon, 17 May 2021 06:03:56 GMT
- Title: Sentence Similarity Based on Contexts
- Authors: Xiaofei Sun, Yuxian Meng, Xiang Ao, Fei Wu, Tianwei Zhang, Jiwei Li
and Chun Fan
- Abstract summary: The proposed framework is based on the core idea that the meaning of a sentence should be defined by its contexts.
It is able to generate high-quality, large-scale dataset with semantic similarity scores between two sentences in an unsupervised manner.
- Score: 31.135984064747607
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Existing methods to measure sentence similarity are faced with two
challenges: (1) labeled datasets are usually limited in size, making them
insufficient to train supervised neural models; (2) there is a training-test
gap for unsupervised language modeling (LM) based models to compute semantic
scores between sentences, since sentence-level semantics are not explicitly
modeled at training. This results in inferior performances in this task. In
this work, we propose a new framework to address these two issues. The proposed
framework is based on the core idea that the meaning of a sentence should be
defined by its contexts, and that sentence similarity can be measured by
comparing the probabilities of generating two sentences given the same context.
The proposed framework is able to generate high-quality, large-scale dataset
with semantic similarity scores between two sentences in an unsupervised
manner, with which the train-test gap can be largely bridged. Extensive
experiments show that the proposed framework achieves significant performance
boosts over existing baselines under both the supervised and unsupervised
settings across different datasets.
Related papers
- DenoSent: A Denoising Objective for Self-Supervised Sentence
Representation Learning [59.4644086610381]
We propose a novel denoising objective that inherits from another perspective, i.e., the intra-sentence perspective.
By introducing both discrete and continuous noise, we generate noisy sentences and then train our model to restore them to their original form.
Our empirical evaluations demonstrate that this approach delivers competitive results on both semantic textual similarity (STS) and a wide range of transfer tasks.
arXiv Detail & Related papers (2024-01-24T17:48:45Z) - Learning to Paraphrase Sentences to Different Complexity Levels [3.0273878903284275]
Sentence simplification is an active research topic in NLP, but its adjacent tasks of sentence complexification and same-level paraphrasing are not.
To train models on all three tasks, we present two new unsupervised datasets.
arXiv Detail & Related papers (2023-08-04T09:43:37Z) - Relational Sentence Embedding for Flexible Semantic Matching [86.21393054423355]
We present Sentence Embedding (RSE), a new paradigm to discover further the potential of sentence embeddings.
RSE is effective and flexible in modeling sentence relations and outperforms a series of state-of-the-art embedding methods.
arXiv Detail & Related papers (2022-12-17T05:25:17Z) - Generate, Discriminate and Contrast: A Semi-Supervised Sentence
Representation Learning Framework [68.04940365847543]
We propose a semi-supervised sentence embedding framework, GenSE, that effectively leverages large-scale unlabeled data.
Our method include three parts: 1) Generate: A generator/discriminator model is jointly trained to synthesize sentence pairs from open-domain unlabeled corpus; 2) Discriminate: Noisy sentence pairs are filtered out by the discriminator to acquire high-quality positive and negative sentence pairs; 3) Contrast: A prompt-based contrastive approach is presented for sentence representation learning with both annotated and synthesized data.
arXiv Detail & Related papers (2022-10-30T10:15:21Z) - Unsupervised Mismatch Localization in Cross-Modal Sequential Data [5.932046800902776]
We develop an unsupervised learning algorithm that can infer the relationship between content-mismatched cross-modal data.
We propose a hierarchical Bayesian deep learning model, named mismatch localization variational autoencoder (ML-VAE), that decomposes the generative process of the speech into hierarchically structured latent variables.
Our experimental results show that ML-VAE successfully locates the mismatch between text and speech, without the need for human annotations.
arXiv Detail & Related papers (2022-05-05T14:23:27Z) - Bridging the Data Gap between Training and Inference for Unsupervised
Neural Machine Translation [49.916963624249355]
A UNMT model is trained on the pseudo parallel data with translated source, and natural source sentences in inference.
The source discrepancy between training and inference hinders the translation performance of UNMT models.
We propose an online self-training approach, which simultaneously uses the pseudo parallel data natural source, translated target to mimic the inference scenario.
arXiv Detail & Related papers (2022-03-16T04:50:27Z) - Retrieve-and-Fill for Scenario-based Task-Oriented Semantic Parsing [110.4684789199555]
We introduce scenario-based semantic parsing: a variant of the original task which first requires disambiguating an utterance's "scenario"
This formulation enables us to isolate coarse-grained and fine-grained aspects of the task, each of which we solve with off-the-shelf neural modules.
Our model is modular, differentiable, interpretable, and allows us to garner extra supervision from scenarios.
arXiv Detail & Related papers (2022-02-02T08:00:21Z) - ConRPG: Paraphrase Generation using Contexts as Regularizer [31.967883219986362]
A long-standing issue with paraphrase generation is how to obtain reliable supervision signals.
We propose an unsupervised paradigm for paraphrase generation based on the assumption that the probabilities of generating two sentences with the same meaning given the same context should be the same.
We propose a pipelined system which consists of paraphrase candidate generation based on contextual language models, candidate filtering using scoring functions, and paraphrase model training based on the selected candidates.
arXiv Detail & Related papers (2021-09-01T12:57:30Z) - Rethinking End-to-End Evaluation of Decomposable Tasks: A Case Study on
Spoken Language Understanding [101.24748444126982]
Decomposable tasks are complex and comprise of a hierarchy of sub-tasks.
Existing benchmarks, however, typically hold out examples for only the surface-level sub-task.
We propose a framework to construct robust test sets using coordinate ascent over sub-task specific utility functions.
arXiv Detail & Related papers (2021-06-29T02:53:59Z) - Reinforcement Learning-based N-ary Cross-Sentence Relation Extraction [3.342376225738321]
Models of n-ary cross sentence relation extraction based on distant supervision assume that consecutive sentences mentioning n entities describe the relation of these n entities.
On the other hand, some non-consecutive sentences also describe one relation and these sentences cannot be labeled under this assumption.
We propose a novel sentence distribution estimator model to address the first problem.
arXiv Detail & Related papers (2020-09-26T20:39:55Z) - SentPWNet: A Unified Sentence Pair Weighting Network for Task-specific
Sentence Embedding [12.020634125787279]
We propose a unified locality weighting and learning framework to learn task-specific sentence embedding.
Our model, SentPWNet, exploits the neighboring spatial distribution of each sentence as locality weight to indicate the informative level of sentence pair.
arXiv Detail & Related papers (2020-05-22T18:32:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.