ConvGQR: Generative Query Reformulation for Conversational Search
- URL: http://arxiv.org/abs/2305.15645v3
- Date: Sat, 27 Jan 2024 17:42:14 GMT
- Title: ConvGQR: Generative Query Reformulation for Conversational Search
- Authors: Fengran Mo, Kelong Mao, Yutao Zhu, Yihong Wu, Kaiyu Huang, Jian-Yun
Nie
- Abstract summary: ConvGQR is a new framework to reformulate conversational queries based on generative pre-trained language models.
We propose a knowledge infusion mechanism to optimize both query reformulation and retrieval.
- Score: 37.54018632257896
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In conversational search, the user's real search intent for the current turn
is dependent on the previous conversation history. It is challenging to
determine a good search query from the whole conversation context. To avoid the
expensive re-training of the query encoder, most existing methods try to learn
a rewriting model to de-contextualize the current query by mimicking the manual
query rewriting. However, manually rewritten queries are not always the best
search queries. Training a rewriting model on them would limit the model's
ability to produce good search queries. Another useful hint is the potential
answer to the question. In this paper, we propose ConvGQR, a new framework to
reformulate conversational queries based on generative pre-trained language
models (PLMs), one for query rewriting and another for generating potential
answers. By combining both, ConvGQR can produce better search queries. In
addition, to relate query reformulation to retrieval performance, we propose a
knowledge infusion mechanism to optimize both query reformulation and
retrieval. Extensive experiments on four conversational search datasets
demonstrate the effectiveness of ConvGQR.
Related papers
- Aligning Query Representation with Rewritten Query and Relevance Judgments in Conversational Search [32.35446999027349]
We leverage both rewritten queries and relevance judgments in the conversational search data to train a better query representation model.
The proposed model -- Query Representation Alignment Conversational Retriever, QRACDR, is tested on eight datasets.
arXiv Detail & Related papers (2024-07-29T17:14:36Z) - Conversational Query Reformulation with the Guidance of Retrieved Documents [4.438698005789677]
We introduce GuideCQR, a framework that refines queries by leveraging key infFormation from the initially retrieved documents.
We show that GuideCQR can get additional performance gains in conversational search using various types of queries, even for queries written by humans.
arXiv Detail & Related papers (2024-07-17T07:39:16Z) - BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval [54.54576644403115]
Many complex real-world queries require in-depth reasoning to identify relevant documents.
We introduce BRIGHT, the first text retrieval benchmark that requires intensive reasoning to retrieve relevant documents.
Our dataset consists of 1,384 real-world queries spanning diverse domains, such as economics, psychology, mathematics, and coding.
arXiv Detail & Related papers (2024-07-16T17:58:27Z) - Adaptive Query Rewriting: Aligning Rewriters through Marginal Probability of Conversational Answers [66.55612528039894]
AdaQR is a framework for training query rewriting models with limited rewrite annotations from seed datasets and completely no passage label.
A novel approach is proposed to assess retriever's preference for these candidates by the probability of answers conditioned on the conversational query.
arXiv Detail & Related papers (2024-06-16T16:09:05Z) - SSP: Self-Supervised Post-training for Conversational Search [63.28684982954115]
We propose fullmodel (model) which is a new post-training paradigm with three self-supervised tasks to efficiently initialize the conversational search model.
To verify the effectiveness of our proposed method, we apply the conversational encoder post-trained by model on the conversational search task using two benchmark datasets: CAsT-19 and CAsT-20.
arXiv Detail & Related papers (2023-07-02T13:36:36Z) - Learning to Relate to Previous Turns in Conversational Search [26.931718474500652]
An effective way to improve retrieval effectiveness is to expand the current query with historical queries.
We propose a new method to select relevant historical queries that are useful for the current query.
arXiv Detail & Related papers (2023-06-05T03:00:10Z) - Decoding a Neural Retriever's Latent Space for Query Suggestion [28.410064376447718]
We show that it is possible to decode a meaningful query from its latent representation and, when moving in the right direction in latent space, to decode a query that retrieves the relevant paragraph.
We employ the query decoder to generate a large synthetic dataset of query reformulations for MSMarco.
On this data, we train a pseudo-relevance feedback (PRF) T5 model for the application of query suggestion.
arXiv Detail & Related papers (2022-10-21T16:19:31Z) - CONQRR: Conversational Query Rewriting for Retrieval with Reinforcement
Learning [16.470428531658232]
We develop a query rewriting model CONQRR that rewrites a conversational question in context into a standalone question.
We show that CONQRR achieves state-of-the-art results on a recent open-domain CQA dataset.
arXiv Detail & Related papers (2021-12-16T01:40:30Z) - Query Resolution for Conversational Search with Limited Supervision [63.131221660019776]
We propose QuReTeC (Query Resolution by Term Classification), a neural query resolution model based on bidirectional transformers.
We show that QuReTeC outperforms state-of-the-art models, and furthermore, that our distant supervision method can be used to substantially reduce the amount of human-curated data required to train QuReTeC.
arXiv Detail & Related papers (2020-05-24T11:37:22Z) - Open-Retrieval Conversational Question Answering [62.11228261293487]
We introduce an open-retrieval conversational question answering (ORConvQA) setting, where we learn to retrieve evidence from a large collection before extracting answers.
We build an end-to-end system for ORConvQA, featuring a retriever, a reranker, and a reader that are all based on Transformers.
arXiv Detail & Related papers (2020-05-22T19:39:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.