Deep Search Query Intent Understanding
- URL: http://arxiv.org/abs/2008.06759v2
- Date: Tue, 18 Aug 2020 04:59:27 GMT
- Title: Deep Search Query Intent Understanding
- Authors: Xiaowei Liu, Weiwei Guo, Huiji Gao, Bo Long
- Abstract summary: This paper aims to provide a comprehensive learning framework for modeling query intent under different stages of a search.
We focus on the design for 1) predicting users' intents as they type in queries on-the-fly in typeahead search using character-level models; and 2) accurate word-level intent prediction models for complete queries.
- Score: 17.79430887321982
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding a user's query intent behind a search is critical for modern
search engine success. Accurate query intent prediction allows the search
engine to better serve the user's need by rendering results from more relevant
categories. This paper aims to provide a comprehensive learning framework for
modeling query intent under different stages of a search. We focus on the
design for 1) predicting users' intents as they type in queries on-the-fly in
typeahead search using character-level models; and 2) accurate word-level
intent prediction models for complete queries. Various deep learning components
for query text understanding are experimented. Offline evaluation and online
A/B test experiments show that the proposed methods are effective in
understanding query intent and efficient to scale for online search systems.
Related papers
- BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval [54.54576644403115]
Many complex real-world queries require in-depth reasoning to identify relevant documents.
We introduce BRIGHT, the first text retrieval benchmark that requires intensive reasoning to retrieve relevant documents.
Our dataset consists of 1,384 real-world queries spanning diverse domains, such as economics, psychology, mathematics, and coding.
arXiv Detail & Related papers (2024-07-16T17:58:27Z) - Query-oriented Data Augmentation for Session Search [71.84678750612754]
We propose query-oriented data augmentation to enrich search logs and empower the modeling.
We generate supplemental training pairs by altering the most important part of a search context.
We develop several strategies to alter the current query, resulting in new training data with varying degrees of difficulty.
arXiv Detail & Related papers (2024-07-04T08:08:33Z) - Evaluating Embedding APIs for Information Retrieval [51.24236853841468]
We evaluate the capabilities of existing semantic embedding APIs on domain generalization and multilingual retrieval.
We find that re-ranking BM25 results using the APIs is a budget-friendly approach and is most effective in English.
For non-English retrieval, re-ranking still improves the results, but a hybrid model with BM25 works best, albeit at a higher cost.
arXiv Detail & Related papers (2023-05-10T16:40:52Z) - Graph Enhanced BERT for Query Understanding [55.90334539898102]
query understanding plays a key role in exploring users' search intents and facilitating users to locate their most desired information.
In recent years, pre-trained language models (PLMs) have advanced various natural language processing tasks.
We propose a novel graph-enhanced pre-training framework, GE-BERT, which can leverage both query content and the query graph.
arXiv Detail & Related papers (2022-04-03T16:50:30Z) - Deep Reinforcement Agent for Efficient Instant Search [14.086339486783018]
We propose to address the load issue by identifying tokens that are semantically more salient towards retrieving relevant documents.
We train a reinforcement agent that interacts directly with the search engine and learns to predict the word's importance.
A novel evaluation framework is presented to study the trade-off between the number of triggered searches and the system's performance.
arXiv Detail & Related papers (2022-03-17T22:47:15Z) - Exposing Query Identification for Search Transparency [69.06545074617685]
We explore the feasibility of approximate exposing query identification (EQI) as a retrieval task by reversing the role of queries and documents in two classes of search systems.
We derive an evaluation metric to measure the quality of a ranking of exposing queries, as well as conducting an empirical analysis focusing on various practical aspects of approximate EQI.
arXiv Detail & Related papers (2021-10-14T20:19:27Z) - Query Understanding via Intent Description Generation [75.64800976586771]
We propose a novel Query-to-Intent-Description (Q2ID) task for query understanding.
Unlike existing ranking tasks which leverage the query and its description to compute the relevance of documents, Q2ID is a reverse task which aims to generate a natural language intent description.
We demonstrate the effectiveness of our model by comparing with several state-of-the-art generation models on the Q2ID task.
arXiv Detail & Related papers (2020-08-25T08:56:40Z) - Query Intent Detection from the SEO Perspective [0.34376560669160383]
We aim to identify the user query's intent by taking advantage of Google results and machine learning methods.
A list of keywords extracted from the clustered queries is used to identify the intent of a new given query.
arXiv Detail & Related papers (2020-06-16T13:08:29Z) - User Intent Inference for Web Search and Conversational Agents [3.9400263964632836]
thesis work focuses on: 1) Utterance topic and intent classification for conversational agents 2) Query intent mining and classification for Web search engines.
To address the first topic, I proposed novel models to incorporate entity information and conversation-context clues to predict both topic and intent of the user's utterances.
For the second research topic, I plan to extend the existing state of the art methods in Web search intent prediction to the e-commerce domain.
arXiv Detail & Related papers (2020-05-28T07:04:42Z) - Modeling Information Need of Users in Search Sessions [5.172625611483604]
We propose a sequence-to-sequence based neural architecture that leverages the set of past queries issued by users.
Firstly, we employ our model for predicting the words in the current query that are important and would be retained in the next query.
We show that our intuitive strategy of capturing information need can yield superior performance at these tasks on two large real-world search log datasets.
arXiv Detail & Related papers (2020-01-03T15:25:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.