HPC Storage Service Autotuning Using Variational-Autoencoder-Guided
Asynchronous Bayesian Optimization
- URL: http://arxiv.org/abs/2210.00798v1
- Date: Mon, 3 Oct 2022 10:12:57 GMT
- Title: HPC Storage Service Autotuning Using Variational-Autoencoder-Guided
Asynchronous Bayesian Optimization
- Authors: Matthieu Dorier, Romain Egele, Prasanna Balaprakash, Jaehoon Koo,
Sandeep Madireddy, Srinivasan Ramesh, Allen D. Malony, Rob Ross
- Abstract summary: We develop a novel variational-autoencoder-guided asynchronous Bayesian optimization method to tune HPC storage service parameters.
We implement our approach within the DeepHyper open-source framework, and apply it to the autotuning of a high-energy physics workflow on Argonne's Theta supercomputer.
Our approach is on par with state-of-the-art autotuning frameworks in speed and outperforms them in resource utilization and parallelization capabilities.
- Score: 3.153934519625761
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Distributed data storage services tailored to specific applications have
grown popular in the high-performance computing (HPC) community as a way to
address I/O and storage challenges. These services offer a variety of specific
interfaces, semantics, and data representations. They also expose many tuning
parameters, making it difficult for their users to find the best configuration
for a given workload and platform.
To address this issue, we develop a novel variational-autoencoder-guided
asynchronous Bayesian optimization method to tune HPC storage service
parameters. Our approach uses transfer learning to leverage prior tuning
results and use a dynamically updated surrogate model to explore the large
parameter search space in a systematic way.
We implement our approach within the DeepHyper open-source framework, and
apply it to the autotuning of a high-energy physics workflow on Argonne's Theta
supercomputer. We show that our transfer-learning approach enables a more than
$40\times$ search speedup over random search, compared with a $2.5\times$ to
$10\times$ speedup when not using transfer learning. Additionally, we show that
our approach is on par with state-of-the-art autotuning frameworks in speed and
outperforms them in resource utilization and parallelization capabilities.
Related papers
- Inference Optimization of Foundation Models on AI Accelerators [68.24450520773688]
Powerful foundation models, including large language models (LLMs), with Transformer architectures have ushered in a new era of Generative AI.
As the number of model parameters reaches to hundreds of billions, their deployment incurs prohibitive inference costs and high latency in real-world scenarios.
This tutorial offers a comprehensive discussion on complementary inference optimization techniques using AI accelerators.
arXiv Detail & Related papers (2024-07-12T09:24:34Z) - Dynamic Adapter Meets Prompt Tuning: Parameter-Efficient Transfer Learning for Point Cloud Analysis [51.14136878142034]
Point cloud analysis has achieved outstanding performance by transferring point cloud pre-trained models.
Existing methods for model adaptation usually update all model parameters, which is inefficient as it relies on high computational costs.
In this paper, we aim to study parameter-efficient transfer learning for point cloud analysis with an ideal trade-off between task performance and parameter efficiency.
arXiv Detail & Related papers (2024-03-03T08:25:04Z) - Towards General and Efficient Online Tuning for Spark [55.30868031221838]
We present a general and efficient Spark tuning framework that can deal with the three issues simultaneously.
We have implemented this framework as an independent cloud service, and applied it to the data platform in Tencent.
arXiv Detail & Related papers (2023-09-05T02:16:45Z) - Hyperparameter Optimization as a Service on INFN Cloud [0.0]
We present a dedicated service based on INFN Cloud to monitor and coordinate multiple training instances, with gradient-less optimization techniques, via simple HTTP requests.
The service, called Hopaas, is made of a web interface and sets of APIs implemented with a FastAPI backend running through Uvicorn and NGINX in a virtual instance of INFN Cloud.
arXiv Detail & Related papers (2023-01-13T12:57:48Z) - Efficient Automated Deep Learning for Time Series Forecasting [42.47842694670572]
We propose an efficient approach for the joint optimization of neural architecture and hyperparameters of the entire data processing pipeline for time series forecasting.
In contrast to common NAS search spaces, we designed a novel neural architecture search space covering various state-of-the-art architectures.
We empirically study several different budget types enabling efficient multi-fidelity optimization on different forecasting datasets.
arXiv Detail & Related papers (2022-05-11T14:03:25Z) - AUTOMATA: Gradient Based Data Subset Selection for Compute-Efficient
Hyper-parameter Tuning [72.54359545547904]
We propose a gradient-based subset selection framework for hyper- parameter tuning.
We show that using gradient-based data subsets for hyper- parameter tuning achieves significantly faster turnaround times and speedups of 3$times$-30$times$.
arXiv Detail & Related papers (2022-03-15T19:25:01Z) - DHA: End-to-End Joint Optimization of Data Augmentation Policy,
Hyper-parameter and Architecture [81.82173855071312]
We propose an end-to-end solution that integrates the AutoML components and returns a ready-to-use model at the end of the search.
Dha achieves state-of-the-art (SOTA) results on various datasets, especially 77.4% accuracy on ImageNet with cell based search space.
arXiv Detail & Related papers (2021-09-13T08:12:50Z) - Amortized Auto-Tuning: Cost-Efficient Transfer Optimization for
Hyperparameter Recommendation [83.85021205445662]
We propose an instantiation--amortized auto-tuning (AT2) to speed up tuning of machine learning models.
We conduct a thorough analysis of the multi-task multi-fidelity Bayesian optimization framework, which leads to the best instantiation--amortized auto-tuning (AT2)
arXiv Detail & Related papers (2021-06-17T00:01:18Z) - A multi-objective perspective on jointly tuning hardware and
hyperparameters [10.605719154114357]
A full AutoML solution requires selecting appropriate hardware automatically.
We adopt a multi-objective approach that selects and adapts the hardware configuration automatically.
We show in extensive NAS and HPO experiments that both ingredients bring significant speed-ups and cost savings.
arXiv Detail & Related papers (2021-06-10T11:52:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.