Generate-and-Retrieve: use your predictions to improve retrieval for
semantic parsing
- URL: http://arxiv.org/abs/2209.14899v1
- Date: Thu, 29 Sep 2022 16:03:29 GMT
- Title: Generate-and-Retrieve: use your predictions to improve retrieval for
semantic parsing
- Authors: Yury Zemlyanskiy, Michiel de Jong, Joshua Ainslie, Panupong Pasupat,
Peter Shaw, Linlu Qiu, Sumit Sanghai, Fei Sha
- Abstract summary: We propose GandR, a retrieval procedure that retrieves exemplars for which outputs are also similar.
GandR first generates a preliminary prediction with input-based retrieval.
Then, it retrieves exemplars with outputs similar to the preliminary prediction which are used to generate a final prediction.
- Score: 25.725176422936766
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: A common recent approach to semantic parsing augments sequence-to-sequence
models by retrieving and appending a set of training samples, called exemplars.
The effectiveness of this recipe is limited by the ability to retrieve
informative exemplars that help produce the correct parse, which is especially
challenging in low-resource settings. Existing retrieval is commonly based on
similarity of query and exemplar inputs. We propose GandR, a retrieval
procedure that retrieves exemplars for which outputs are also similar.
GandRfirst generates a preliminary prediction with input-based retrieval. Then,
it retrieves exemplars with outputs similar to the preliminary prediction which
are used to generate a final prediction. GandR sets the state of the art on
multiple low-resource semantic parsing tasks.
Related papers
- End-to-End Trainable Retrieval-Augmented Generation for Relation Extraction [7.613942320502336]
We propose a novel End-to-end Trainable Retrieval-Augmented Generation (ETRAG)
ETRAG allows end-to-end optimization of the entire model, including the retriever, for the relation extraction objective.
We evaluate the relation extraction performance of ETRAG on the TACRED dataset, which is a standard benchmark for relation extraction.
arXiv Detail & Related papers (2024-06-06T07:01:50Z) - Adapting Dual-encoder Vision-language Models for Paraphrased Retrieval [55.90407811819347]
We consider the task of paraphrased text-to-image retrieval where a model aims to return similar results given a pair of paraphrased queries.
We train a dual-encoder model starting from a language model pretrained on a large text corpus.
Compared to public dual-encoder models such as CLIP and OpenCLIP, the model trained with our best adaptation strategy achieves a significantly higher ranking similarity for paraphrased queries.
arXiv Detail & Related papers (2024-05-06T06:30:17Z) - Generative Retrieval as Multi-Vector Dense Retrieval [71.75503049199897]
Generative retrieval generates identifiers of relevant documents in an end-to-end manner.
Prior work has demonstrated that generative retrieval with atomic identifiers is equivalent to single-vector dense retrieval.
We show that generative retrieval and multi-vector dense retrieval share the same framework for measuring the relevance to a query of a document.
arXiv Detail & Related papers (2024-03-31T13:29:43Z) - Dense X Retrieval: What Retrieval Granularity Should We Use? [56.90827473115201]
Often-overlooked design choice is the retrieval unit in which the corpus is indexed, e.g. document, passage, or sentence.
We introduce a novel retrieval unit, proposition, for dense retrieval.
Experiments reveal that indexing a corpus by fine-grained units such as propositions significantly outperforms passage-level units in retrieval tasks.
arXiv Detail & Related papers (2023-12-11T18:57:35Z) - ReFIT: Relevance Feedback from a Reranker during Inference [109.33278799999582]
Retrieve-and-rerank is a prevalent framework in neural information retrieval.
We propose to leverage the reranker to improve recall by making it provide relevance feedback to the retriever at inference time.
arXiv Detail & Related papers (2023-05-19T15:30:33Z) - Recommender Systems with Generative Retrieval [58.454606442670034]
We propose a novel generative retrieval approach, where the retrieval model autoregressively decodes the identifiers of the target candidates.
To that end, we create semantically meaningful of codewords to serve as a Semantic ID for each item.
We show that recommender systems trained with the proposed paradigm significantly outperform the current SOTA models on various datasets.
arXiv Detail & Related papers (2023-05-08T21:48:17Z) - Controllable Semantic Parsing via Retrieval Augmentation [14.528396278058285]
We propose ControllAble Semantic generative model via Exemplar Retrieval (CASPER)
We show that CASPER can parse queries in a new domain, adapt the prediction toward the specified patterns, or adapt to new semantic schemas without having to further re-train the model.
arXiv Detail & Related papers (2021-10-16T03:34:49Z) - Graph-Based Decoding for Task Oriented Semantic Parsing [16.054030490095464]
We formulate semantic parsing as a dependency parsing task, applying graph-based decoding techniques developed for syntactic parsing.
We find that our graph-based approach is competitive with sequence decoders on the standard setting, and offers significant improvements in data efficiency and settings where partially-annotated data is available.
arXiv Detail & Related papers (2021-09-09T23:22:09Z) - Generation-Augmented Retrieval for Open-domain Question Answering [134.27768711201202]
Generation-Augmented Retrieval (GAR) for answering open-domain questions.
We show that generating diverse contexts for a query is beneficial as fusing their results consistently yields better retrieval accuracy.
GAR achieves state-of-the-art performance on Natural Questions and TriviaQA datasets under the extractive QA setup when equipped with an extractive reader.
arXiv Detail & Related papers (2020-09-17T23:08:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.