Utilizing deep learning for automated tuning of database management
systems
- URL: http://arxiv.org/abs/2306.14349v1
- Date: Sun, 25 Jun 2023 21:50:14 GMT
- Title: Utilizing deep learning for automated tuning of database management
systems
- Authors: Karthick Prasad Gunasekaran, Kajal Tiwari, Rachana Acharya
- Abstract summary: OtterTune identifies influential knobs, analyze previously unseen workloads, and provide recommendations for knob settings.
The effectiveness of this approach is demonstrated through the evaluation of a new tool called OtterTune on three different database management systems (DBMSs)
- Score: 0.12891210250935148
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Managing the configurations of a database system poses significant challenges
due to the multitude of configuration knobs that impact various system
aspects.The lack of standardization, independence, and universality among these
knobs further complicates the task of determining the optimal settings.To
address this issue, an automated solution leveraging supervised and
unsupervised machine learning techniques was developed.This solution aims to
identify influential knobs, analyze previously unseen workloads, and provide
recommendations for knob settings.The effectiveness of this approach is
demonstrated through the evaluation of a new tool called OtterTune [1] on three
different database management systems (DBMSs).The results indicate that
OtterTune's recommendations are comparable to or even surpass the
configurations generated by existing tools or human experts.In this study, we
build upon the automated technique introduced in the original OtterTune paper,
utilizing previously collected training data to optimize new DBMS
deployments.By employing supervised and unsupervised machine learning methods,
we focus on improving latency prediction.Our approach expands upon the methods
proposed in the original paper by incorporating GMM clustering to streamline
metrics selection and combining ensemble models (such as RandomForest) with
non-linear models (like neural networks) for more accurate prediction modeling.
Related papers
- Federated Learning with Projected Trajectory Regularization [65.6266768678291]
Federated learning enables joint training of machine learning models from distributed clients without sharing their local data.
One key challenge in federated learning is to handle non-identically distributed data across the clients.
We propose a novel federated learning framework with projected trajectory regularization (FedPTR) for tackling the data issue.
arXiv Detail & Related papers (2023-12-22T02:12:08Z) - End-to-End Meta-Bayesian Optimisation with Transformer Neural Processes [52.818579746354665]
This paper proposes the first end-to-end differentiable meta-BO framework that generalises neural processes to learn acquisition functions via transformer architectures.
We enable this end-to-end framework with reinforcement learning (RL) to tackle the lack of labelled acquisition data.
arXiv Detail & Related papers (2023-05-25T10:58:46Z) - Deep learning based Auto Tuning for Database Management System [0.12891210250935148]
In this work, we extend an automated technique based on Ottertune to reuse data gathered from previous sessions to tune new deployments with the help of supervised and unsupervised machine learning methods to improve latency prediction.
We use GMM clustering to prune metrics and combine ensemble models, such as RandomForest, with non-linear models, like neural networks, for prediction modeling.
arXiv Detail & Related papers (2023-04-25T11:52:52Z) - Optimizing Closed-Loop Performance with Data from Similar Systems: A
Bayesian Meta-Learning Approach [1.370633147306388]
We propose the use of meta-learning to generate an initial surrogate model based on data collected from performance optimization tasks.
The effectiveness of our proposed DKN-BO approach for speeding up control system performance optimization is demonstrated.
arXiv Detail & Related papers (2022-10-31T18:25:47Z) - Consolidated learning -- a domain-specific model-free optimization
strategy with examples for XGBoost and MIMIC-IV [4.370097023410272]
This paper proposes a new formulation of the tuning problem, called consolidated learning.
In such settings, we are interested in the total optimization time rather than tuning for a single task.
We demonstrate the effectiveness of this approach through an empirical study for XGBoost algorithm and the collection of predictive tasks extracted from the MIMIC-IV medical database.
arXiv Detail & Related papers (2022-01-27T21:38:53Z) - Top-N Recommendation with Counterfactual User Preference Simulation [26.597102553608348]
Top-N recommendation, which aims to learn user ranking-based preference, has long been a fundamental problem in a wide range of applications.
In this paper, we propose to reformulate the recommendation task within the causal inference framework to handle the data scarce problem.
arXiv Detail & Related papers (2021-09-02T14:28:46Z) - Deep Variational Models for Collaborative Filtering-based Recommender
Systems [63.995130144110156]
Deep learning provides accurate collaborative filtering models to improve recommender system results.
Our proposed models apply the variational concept to injectity in the latent space of the deep architecture.
Results show the superiority of the proposed approach in scenarios where the variational enrichment exceeds the injected noise effect.
arXiv Detail & Related papers (2021-07-27T08:59:39Z) - Hybrid Model with Time Modeling for Sequential Recommender Systems [0.15229257192293202]
Booking.com organized the WSDM WebTour 2021 Challenge, which aims to benchmark models to recommend the final city in a trip.
We conducted several experiments to test different state-of-the-art deep learning architectures for recommender systems.
Our experimental result shows that the improved NARM outperforms all other state-of-the-art benchmark methods.
arXiv Detail & Related papers (2021-03-07T19:28:22Z) - Few-Shot Named Entity Recognition: A Comprehensive Study [92.40991050806544]
We investigate three schemes to improve the model generalization ability for few-shot settings.
We perform empirical comparisons on 10 public NER datasets with various proportions of labeled data.
We create new state-of-the-art results on both few-shot and training-free settings.
arXiv Detail & Related papers (2020-12-29T23:43:16Z) - DAGA: Data Augmentation with a Generation Approach for Low-resource
Tagging Tasks [88.62288327934499]
We propose a novel augmentation method with language models trained on the linearized labeled sentences.
Our method is applicable to both supervised and semi-supervised settings.
arXiv Detail & Related papers (2020-11-03T07:49:15Z) - BREEDS: Benchmarks for Subpopulation Shift [98.90314444545204]
We develop a methodology for assessing the robustness of models to subpopulation shift.
We leverage the class structure underlying existing datasets to control the data subpopulations that comprise the training and test distributions.
Applying this methodology to the ImageNet dataset, we create a suite of subpopulation shift benchmarks of varying granularity.
arXiv Detail & Related papers (2020-08-11T17:04:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.