Approximating Aggregated SQL Queries With LSTM Networks
- URL: http://arxiv.org/abs/2010.13149v3
- Date: Tue, 5 Jan 2021 11:32:09 GMT
- Title: Approximating Aggregated SQL Queries With LSTM Networks
- Authors: Nir Regev, Lior Rokach, Asaf Shabtai
- Abstract summary: We present a method for query approximation, also known as approximate query processing (AQP)
We use LSTM network to learn the relationship between queries and their results, and to provide a rapid inference layer for predicting query results.
Our method was able to predict up to 120,000 queries in a second, and with a single query latency of no more than 2ms.
- Score: 31.528524004435933
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Despite continuous investments in data technologies, the latency of querying
data still poses a significant challenge. Modern analytic solutions require
near real-time responsiveness both to make them interactive and to support
automated processing. Current technologies (Hadoop, Spark, Dataflow) scan the
dataset to execute queries. They focus on providing a scalable data storage to
maximize task execution speed. We argue that these solutions fail to offer an
adequate level of interactivity since they depend on continual access to data.
In this paper we present a method for query approximation, also known as
approximate query processing (AQP), that reduce the need to scan data during
inference (query calculation), thus enabling a rapid query processing tool. We
use LSTM network to learn the relationship between queries and their results,
and to provide a rapid inference layer for predicting query results. Our method
(referred as ``Hunch``) produces a lightweight LSTM network which provides a
high query throughput. We evaluated our method using twelve datasets and
compared to state-of-the-art AQP engines (VerdictDB, BlinkDB) from query
latency, model weight and accuracy perspectives. The results show that our
method predicted queries' results with a normalized root mean squared error
(NRMSE) ranging from approximately 1\% to 4\% which in the majority of our data
sets was better then the compared benchmarks. Moreover, our method was able to
predict up to 120,000 queries in a second (streamed together), and with a
single query latency of no more than 2ms.
Related papers
- Enhancing Long Context Performance in LLMs Through Inner Loop Query Mechanism [2.919891871101241]
Transformers have a quadratic scaling of computational complexity with input size.
Retrieval-augmented generation (RAG) can better handle longer contexts by using a retrieval system.
We introduce a novel approach, Inner Loop Memory Augmented Tree Retrieval (ILM-TR)
arXiv Detail & Related papers (2024-10-11T19:49:05Z) - UQE: A Query Engine for Unstructured Databases [71.49289088592842]
We investigate the potential of Large Language Models to enable unstructured data analytics.
We propose a new Universal Query Engine (UQE) that directly interrogates and draws insights from unstructured data collections.
arXiv Detail & Related papers (2024-06-23T06:58:55Z) - User Intent Recognition and Semantic Cache Optimization-Based Query Processing Framework using CFLIS and MGR-LAU [0.0]
This work analyzed the informational, navigational, and transactional-based intents in queries for enhanced QP.
For efficient QP, the data is structured using Epanechnikov Kernel-Ordering Points To Identify the Clustering Structure (EK-OPTICS)
The extracted features, detected intents and structured data are inputted to the Multi-head Gated Recurrent Learnable Attention Unit (MGR-LAU)
arXiv Detail & Related papers (2024-06-06T20:28:05Z) - Optimizing LLM Queries in Relational Workloads [58.254894049950366]
We show how to optimize Large Language Models (LLMs) inference for analytical workloads that invoke LLMs within relational queries.
We implement these optimizations in Apache Spark, with vLLM as the model serving backend.
We achieve up to 4.4x improvement in end-to-end latency on a benchmark of diverse LLM-based queries on real datasets.
arXiv Detail & Related papers (2024-03-09T07:01:44Z) - MeaeQ: Mount Model Extraction Attacks with Efficient Queries [6.1106195466129485]
We study model extraction attacks in natural language processing (NLP)
We propose MeaeQ, a straightforward yet effective method to address these issues.
MeaeQ achieves higher functional similarity to the victim model than baselines while requiring fewer queries.
arXiv Detail & Related papers (2023-10-21T16:07:16Z) - Improving Text Matching in E-Commerce Search with A Rationalizable,
Intervenable and Fast Entity-Based Relevance Model [78.80174696043021]
We propose a novel model called the Entity-Based Relevance Model (EBRM)
The decomposition allows us to use a Cross-encoder QE relevance module for high accuracy.
We also show that pretraining the QE module with auto-generated QE data from user logs can further improve the overall performance.
arXiv Detail & Related papers (2023-07-01T15:44:53Z) - Kepler: Robust Learning for Faster Parametric Query Optimization [5.6119420695093245]
We propose an end-to-end learning-based approach to parametric query optimization.
Kepler achieves significant improvements in query runtime on multiple datasets.
arXiv Detail & Related papers (2023-06-11T22:39:28Z) - On Efficient Approximate Queries over Machine Learning Models [30.26180913049285]
We develop a novel unified framework for approximate query answering by leveraging a proxy to minimize the oracle usage.
Our framework uses a judicious combination of invoking the expensive oracle on data samples and applying the cheap proxy on the objects in the DB.
Our algorithms outperform the state-of-the-art and achieve high result quality with provable statistical guarantees.
arXiv Detail & Related papers (2022-06-06T18:35:19Z) - IRLI: Iterative Re-partitioning for Learning to Index [104.72641345738425]
Methods have to trade between obtaining high accuracy while maintaining load balance and scalability in distributed settings.
We propose a novel approach called IRLI, which iteratively partitions the items by learning the relevant buckets directly from the query-item relevance data.
We mathematically show that IRLI retrieves the correct item with high probability under very natural assumptions and provides superior load balancing.
arXiv Detail & Related papers (2021-03-17T23:13:25Z) - Data Augmentation for Abstractive Query-Focused Multi-Document
Summarization [129.96147867496205]
We present two QMDS training datasets, which we construct using two data augmentation methods.
These two datasets have complementary properties, i.e., QMDSCNN has real summaries but queries are simulated, while QMDSIR has real queries but simulated summaries.
We build end-to-end neural network models on the combined datasets that yield new state-of-the-art transfer results on DUC datasets.
arXiv Detail & Related papers (2021-03-02T16:57:01Z) - Query Resolution for Conversational Search with Limited Supervision [63.131221660019776]
We propose QuReTeC (Query Resolution by Term Classification), a neural query resolution model based on bidirectional transformers.
We show that QuReTeC outperforms state-of-the-art models, and furthermore, that our distant supervision method can be used to substantially reduce the amount of human-curated data required to train QuReTeC.
arXiv Detail & Related papers (2020-05-24T11:37:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.