Session-Aware Query Auto-completion using Extreme Multi-label Ranking
- URL: http://arxiv.org/abs/2012.07654v1
- Date: Wed, 9 Dec 2020 17:56:22 GMT
- Title: Session-Aware Query Auto-completion using Extreme Multi-label Ranking
- Authors: Nishant Yadav, Rajat Sen, Daniel N. Hill, Arya Mazumdar, Inderjit S.
Dhillon
- Abstract summary: We take the novel approach of modeling session-aware query auto-completion as an e Multi-Xtreme Ranking (XMR) problem.
We adapt a popular XMR algorithm for this purpose by proposing several modifications to the key steps in the algorithm.
Our approach meets the stringent latency requirements for auto-complete systems while leveraging session information in making suggestions.
- Score: 61.753713147852125
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Query auto-completion is a fundamental feature in search engines where the
task is to suggest plausible completions of a prefix typed in the search bar.
Previous queries in the user session can provide useful context for the user's
intent and can be leveraged to suggest auto-completions that are more relevant
while adhering to the user's prefix. Such session-aware query auto-completions
can be generated by sequence-to-sequence models; however, these generative
approaches often do not meet the stringent latency requirements of responding
to each user keystroke. Moreover, there is a danger of showing non-sensical
queries in a generative approach. Another solution is to pre-compute a
relatively small subset of relevant queries for common prefixes and rank them
based on the context. However, such an approach would fail if no relevant
queries for the current context are present in the pre-computed set.
In this paper, we provide a solution to this problem: we take the novel
approach of modeling session-aware query auto-completion as an eXtreme
Multi-Label Ranking (XMR) problem where the input is the previous query in the
session and the user's current prefix, while the output space is the set of
millions of queries entered by users in the recent past. We adapt a popular XMR
algorithm for this purpose by proposing several modifications to the key steps
in the algorithm. The proposed modifications yield a 230% improvement in terms
of Mean Reciprocal Rank over the baseline XMR approach on a public search logs
dataset. Our approach meets the stringent latency requirements for
auto-complete systems while leveraging session information in making
suggestions. We show that session context leads to significant improvements in
the quality of query auto-completions; in particular, for short prefixes with
up to 3 characters, we see a 32% improvement over baselines that meet latency
requirements.
Related papers
- Database-Augmented Query Representation for Information Retrieval [59.57065228857247]
We present a novel retrieval framework called Database-Augmented Query representation (DAQu)
DAQu augments the original query with various (query-related) metadata across multiple tables.
We validate DAQu in diverse retrieval scenarios that can incorporate metadata from the relational database.
arXiv Detail & Related papers (2024-06-23T05:02:21Z) - Adaptive Query Rewriting: Aligning Rewriters through Marginal Probability of Conversational Answers [66.55612528039894]
AdaQR is a framework for training query rewriting models with limited rewrite annotations from seed datasets and completely no passage label.
A novel approach is proposed to assess retriever's preference for these candidates by the probability of answers conditioned on the conversational query.
arXiv Detail & Related papers (2024-06-16T16:09:05Z) - User Intent Recognition and Semantic Cache Optimization-Based Query Processing Framework using CFLIS and MGR-LAU [0.0]
This work analyzed the informational, navigational, and transactional-based intents in queries for enhanced QP.
For efficient QP, the data is structured using Epanechnikov Kernel-Ordering Points To Identify the Clustering Structure (EK-OPTICS)
The extracted features, detected intents and structured data are inputted to the Multi-head Gated Recurrent Learnable Attention Unit (MGR-LAU)
arXiv Detail & Related papers (2024-06-06T20:28:05Z) - Ask Optimal Questions: Aligning Large Language Models with Retriever's
Preference in Conversational Search [25.16282868262589]
RetPO is designed to optimize a language model (LM) for reformulating search queries in line with the preferences of the target retrieval systems.
We construct a large-scale dataset called Retrievers' Feedback on over 410K query rewrites across 12K conversations.
The resulting model achieves state-of-the-art performance on two recent conversational search benchmarks.
arXiv Detail & Related papers (2024-02-19T04:41:31Z) - Recommender Systems with Generative Retrieval [58.454606442670034]
We propose a novel generative retrieval approach, where the retrieval model autoregressively decodes the identifiers of the target candidates.
To that end, we create semantically meaningful of codewords to serve as a Semantic ID for each item.
We show that recommender systems trained with the proposed paradigm significantly outperform the current SOTA models on various datasets.
arXiv Detail & Related papers (2023-05-08T21:48:17Z) - Context-Aware Query Rewriting for Improving Users' Search Experience on
E-commerce Websites [47.04727122209316]
E-commerce queries are often short and ambiguous.
Users tend to enter multiple searches, which we call context, before purchasing.
We propose an end-to-end context-aware query rewriting model.
arXiv Detail & Related papers (2022-09-15T19:46:01Z) - Efficient Neural Query Auto Completion [17.58784759652327]
Three major challenges are observed for a query auto completion system.
Traditional QAC systems rely on handcrafted features such as the query candidate frequency in search logs.
We propose an efficient neural QAC system with effective context modeling to overcome these challenges.
arXiv Detail & Related papers (2020-08-06T21:28:36Z) - Query Resolution for Conversational Search with Limited Supervision [63.131221660019776]
We propose QuReTeC (Query Resolution by Term Classification), a neural query resolution model based on bidirectional transformers.
We show that QuReTeC outperforms state-of-the-art models, and furthermore, that our distant supervision method can be used to substantially reduce the amount of human-curated data required to train QuReTeC.
arXiv Detail & Related papers (2020-05-24T11:37:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.