Frugal Optimization for Cost-related Hyperparameters
- URL: http://arxiv.org/abs/2005.01571v3
- Date: Tue, 22 Dec 2020 20:48:40 GMT
- Title: Frugal Optimization for Cost-related Hyperparameters
- Authors: Qingyun Wu, Chi Wang, Silu Huang
- Abstract summary: We develop a new cost-frugal HPO solution for machine learning algorithms.
We prove a convergence rate of $O(fracsqrtdsqrtK)$ and an $O(depsilon-2)$-approximation guarantee on the total cost.
We provide strong empirical results in comparison with state-of-the-art HPO methods on large AutoML benchmarks.
- Score: 43.599155206275306
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The increasing demand for democratizing machine learning algorithms calls for
hyperparameter optimization (HPO) solutions at low cost. Many machine learning
algorithms have hyperparameters which can cause a large variation in the
training cost. But this effect is largely ignored in existing HPO methods,
which are incapable to properly control cost during the optimization process.
To address this problem, we develop a new cost-frugal HPO solution. The core of
our solution is a simple but new randomized direct-search method, for which we
prove a convergence rate of $O(\frac{\sqrt{d}}{\sqrt{K}})$ and an
$O(d\epsilon^{-2})$-approximation guarantee on the total cost. We provide
strong empirical results in comparison with state-of-the-art HPO methods on
large AutoML benchmarks.
Related papers
- Time-Varying Convex Optimization with $O(n)$ Computational Complexity [0.0]
We consider the problem of unconstrained convex optimization, where the cost function changes with time.
Our proposed algorithms only require the first-order derivatives of the cost function with respect to the decision variable.
Specifically, the proposed algorithms reduce computational cost from $(n3)$ to $O(n)$ per timestep.
arXiv Detail & Related papers (2024-10-19T06:45:05Z) - Reinforcement Learning from Human Feedback with Active Queries [67.27150911254155]
Current reinforcement learning approaches often require a large amount of human-labelled preference data.
We propose query-efficient RLHF methods, inspired by the success of active learning.
Our experiments show that ADPO, while only making about half of queries for human preference, matches the performance of the state-of-the-art DPO method.
arXiv Detail & Related papers (2024-02-14T18:58:40Z) - A New Linear Scaling Rule for Private Adaptive Hyperparameter Optimization [57.450449884166346]
We propose an adaptive HPO method to account for the privacy cost of HPO.
We obtain state-of-the-art performance on 22 benchmark tasks, across computer vision and natural language processing, across pretraining and finetuning.
arXiv Detail & Related papers (2022-12-08T18:56:37Z) - Breaking the Sample Complexity Barrier to Regret-Optimal Model-Free
Reinforcement Learning [52.76230802067506]
A novel model-free algorithm is proposed to minimize regret in episodic reinforcement learning.
The proposed algorithm employs an em early-settled reference update rule, with the aid of two Q-learning sequences.
The design principle of our early-settled variance reduction method might be of independent interest to other RL settings.
arXiv Detail & Related papers (2021-10-09T21:13:48Z) - A Nonmyopic Approach to Cost-Constrained Bayesian Optimization [10.078368988372247]
We formulate cost-constrained BO as a constrained Markov decision process (CMDP)
We develop an efficient rollout approximation to the optimal CMDP policy that takes both the cost and future iterations into account.
arXiv Detail & Related papers (2021-06-10T22:44:37Z) - Cost-Efficient Online Hyperparameter Optimization [94.60924644778558]
We propose an online HPO algorithm that reaches human expert-level performance within a single run of the experiment.
Our proposed online HPO algorithm reaches human expert-level performance within a single run of the experiment, while incurring only modest computational overhead compared to regular training.
arXiv Detail & Related papers (2021-01-17T04:55:30Z) - Efficient Automatic CASH via Rising Bandits [37.09843193057032]
We propose an alternating optimization framework for CASH problem.
We also introduce Rising Bandits, a CASH-oriented Multi-Armed Bandits (MAB) variant, to model the algorithm selection in CASH.
This framework can take the advantages of both BO in solving the HPO problem and the MABs in accelerating the algorithm selection.
arXiv Detail & Related papers (2020-12-08T11:29:57Z) - Combining Deep Learning and Optimization for Security-Constrained
Optimal Power Flow [94.24763814458686]
Security-constrained optimal power flow (SCOPF) is fundamental in power systems.
Modeling of APR within the SCOPF problem results in complex large-scale mixed-integer programs.
This paper proposes a novel approach that combines deep learning and robust optimization techniques.
arXiv Detail & Related papers (2020-07-14T12:38:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.