Towards General and Efficient Online Tuning for Spark
- URL: http://arxiv.org/abs/2309.01901v1
- Date: Tue, 5 Sep 2023 02:16:45 GMT
- Title: Towards General and Efficient Online Tuning for Spark
- Authors: Yang Li, Huaijun Jiang, Yu Shen, Yide Fang, Xiaofeng Yang, Danqing
Huang, Xinyi Zhang, Wentao Zhang, Ce Zhang, Peng Chen and Bin Cui
- Abstract summary: We present a general and efficient Spark tuning framework that can deal with the three issues simultaneously.
We have implemented this framework as an independent cloud service, and applied it to the data platform in Tencent.
- Score: 55.30868031221838
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The distributed data analytic system -- Spark is a common choice for
processing massive volumes of heterogeneous data, while it is challenging to
tune its parameters to achieve high performance. Recent studies try to employ
auto-tuning techniques to solve this problem but suffer from three issues:
limited functionality, high overhead, and inefficient search.
In this paper, we present a general and efficient Spark tuning framework that
can deal with the three issues simultaneously. First, we introduce a
generalized tuning formulation, which can support multiple tuning goals and
constraints conveniently, and a Bayesian optimization (BO) based solution to
solve this generalized optimization problem. Second, to avoid high overhead
from additional offline evaluations in existing methods, we propose to tune
parameters along with the actual periodic executions of each job (i.e., online
evaluations). To ensure safety during online job executions, we design a safe
configuration acquisition method that models the safe region. Finally, three
innovative techniques are leveraged to further accelerate the search process:
adaptive sub-space generation, approximate gradient descent, and meta-learning
method.
We have implemented this framework as an independent cloud service, and
applied it to the data platform in Tencent. The empirical results on both
public benchmarks and large-scale production tasks demonstrate its superiority
in terms of practicality, generality, and efficiency. Notably, this service
saves an average of 57.00% memory cost and 34.93% CPU cost on 25K in-production
tasks within 20 iterations, respectively.
Related papers
- Efficient Federated Learning Using Dynamic Update and Adaptive Pruning with Momentum on Shared Server Data [59.6985168241067]
Federated Learning (FL) encounters two important problems, i.e., low training efficiency and limited computational resources.
We propose a new FL framework, FedDUMAP, to leverage the shared insensitive data on the server and the distributed data in edge devices.
Our proposed FL model, FedDUMAP, combines the three original techniques and has a significantly better performance compared with baseline approaches.
arXiv Detail & Related papers (2024-08-11T02:59:11Z) - Controllable Prompt Tuning For Balancing Group Distributional Robustness [53.336515056479705]
We introduce an optimization scheme to achieve good performance across groups and find a good solution for all without severely sacrificing performance on any of them.
We propose Controllable Prompt Tuning (CPT), which couples our approach with prompt-tuning techniques.
On spurious correlation benchmarks, our procedures achieve state-of-the-art results across both transformer and non-transformer architectures, as well as unimodal and multimodal data.
arXiv Detail & Related papers (2024-03-05T06:23:55Z) - Federated Learning of Large Language Models with Parameter-Efficient
Prompt Tuning and Adaptive Optimization [71.87335804334616]
Federated learning (FL) is a promising paradigm to enable collaborative model training with decentralized data.
The training process of Large Language Models (LLMs) generally incurs the update of significant parameters.
This paper proposes an efficient partial prompt tuning approach to improve performance and efficiency simultaneously.
arXiv Detail & Related papers (2023-10-23T16:37:59Z) - Score-Based Methods for Discrete Optimization in Deep Learning [30.446056972242616]
We investigate a score-based approximation framework to solve such problems.
We experimentally demonstrate, in adversarial set classification tasks, that our method achieves a superior trade-off in terms of speed and solution quality compared to methods.
arXiv Detail & Related papers (2023-10-15T17:14:17Z) - Large-Batch, Iteration-Efficient Neural Bayesian Design Optimization [37.339567743948955]
We present a novel Bayesian optimization framework specifically tailored to address the limitations of BO.
Our key contribution is a highly scalable, sample-based acquisition function that performs a non-dominated sorting of objectives.
We show that our acquisition function in combination with different Bayesian neural network surrogates is effective in data-intensive environments with a minimal number of iterations.
arXiv Detail & Related papers (2023-06-01T19:10:57Z) - Learning to Optimize Permutation Flow Shop Scheduling via Graph-based
Imitation Learning [70.65666982566655]
Permutation flow shop scheduling (PFSS) is widely used in manufacturing systems.
We propose to train the model via expert-driven imitation learning, which accelerates convergence more stably and accurately.
Our model's network parameters are reduced to only 37% of theirs, and the solution gap of our model towards the expert solutions decreases from 6.8% to 1.3% on average.
arXiv Detail & Related papers (2022-10-31T09:46:26Z) - TCT: Convexifying Federated Learning using Bootstrapped Neural Tangent
Kernels [141.29156234353133]
State-of-the-art convex learning methods can perform far worse than their centralized counterparts when clients have dissimilar data distributions.
We show this disparity can largely be attributed to challenges presented by non-NISTity.
We propose a Train-Convexify neural network (TCT) procedure to sidestep this issue.
arXiv Detail & Related papers (2022-07-13T16:58:22Z) - Efficient Hyperparameter Tuning for Large Scale Kernel Ridge Regression [19.401624974011746]
We propose a complexity regularization criterion based on a data dependent penalty, and discuss its efficient optimization.
Our analysis shows the benefit of the proposed approach, that we hence incorporate in a library for large scale kernel methods to derive adaptively tuned solutions.
arXiv Detail & Related papers (2022-01-17T09:57:32Z) - DeLag: Using Multi-Objective Optimization to Enhance the Detection of
Latency Degradation Patterns in Service-based Systems [0.76146285961466]
We present DeLag, a novel automated search-based approach for diagnosing performance issues in service-based systems.
DeLag simultaneously searches for multiple latency patterns while optimizing precision, recall and dissimilarity.
arXiv Detail & Related papers (2021-10-21T13:59:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.