Boosting One-Point Derivative-Free Online Optimization via Residual
Feedback
- URL: http://arxiv.org/abs/2010.07378v3
- Date: Thu, 3 Dec 2020 02:35:11 GMT
- Title: Boosting One-Point Derivative-Free Online Optimization via Residual
Feedback
- Authors: Yan Zhang, Yi Zhou, Kaiyi Ji, Michael M. Zavlanos
- Abstract summary: We propose a new one-point feedback method for online optimization.
We show that using residual feedback can produce much smaller variance compared to conventional one-point feedback methods.
Our regret bounds are tighter compared to existing regret bounds for ZO with conventional one-point feedback.
- Score: 28.679167097106813
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Zeroth-order optimization (ZO) typically relies on two-point feedback to
estimate the unknown gradient of the objective function. Nevertheless,
two-point feedback can not be used for online optimization of time-varying
objective functions, where only a single query of the function value is
possible at each time step. In this work, we propose a new one-point feedback
method for online optimization that estimates the objective function gradient
using the residual between two feedback points at consecutive time instants.
Moreover, we develop regret bounds for ZO with residual feedback for both
convex and nonconvex online optimization problems. Specifically, for both
deterministic and stochastic problems and for both Lipschitz and smooth
objective functions, we show that using residual feedback can produce gradient
estimates with much smaller variance compared to conventional one-point
feedback methods. As a result, our regret bounds are much tighter compared to
existing regret bounds for ZO with conventional one-point feedback, which
suggests that ZO with residual feedback can better track the optimizer of
online optimization problems. Additionally, our regret bounds rely on weaker
assumptions than those used in conventional one-point feedback methods.
Numerical experiments show that ZO with residual feedback significantly
outperforms existing one-point feedback methods also in practice.
Related papers
- Stochastic Zeroth-Order Optimization under Strongly Convexity and Lipschitz Hessian: Minimax Sample Complexity [59.75300530380427]
We consider the problem of optimizing second-order smooth and strongly convex functions where the algorithm is only accessible to noisy evaluations of the objective function it queries.
We provide the first tight characterization for the rate of the minimax simple regret by developing matching upper and lower bounds.
arXiv Detail & Related papers (2024-06-28T02:56:22Z) - Principled Preferential Bayesian Optimization [22.269732173306192]
We study the problem of preferential Bayesian optimization (BO)
We aim to optimize a black-box function with only preference feedback over a pair of candidate solutions.
An optimistic algorithm with an efficient computational method is then developed to solve the problem.
arXiv Detail & Related papers (2024-02-08T02:57:47Z) - Faster Convergence with Multiway Preferences [99.68922143784306]
We consider the sign-function-based comparison feedback model and analyze the convergence rates with batched and multiway comparisons.
Our work is the first to study the problem of convex optimization with multiway preferences and analyze the optimal convergence rates.
arXiv Detail & Related papers (2023-12-19T01:52:13Z) - Posterior Sampling with Delayed Feedback for Reinforcement Learning with
Linear Function Approximation [62.969796245827006]
Delayed-PSVI is an optimistic value-based algorithm that explores the value function space via noise perturbation with posterior sampling.
We show our algorithm achieves $widetildeO(sqrtd3H3 T + d2H2 E[tau]$ worst-case regret in the presence of unknown delays.
We incorporate a gradient-based approximate sampling scheme via Langevin dynamics for Delayed-LPSVI.
arXiv Detail & Related papers (2023-10-29T06:12:43Z) - Vanishing Point Estimation in Uncalibrated Images with Prior Gravity
Direction [82.72686460985297]
We tackle the problem of estimating a Manhattan frame.
We derive two new 2-line solvers, one of which does not suffer from singularities affecting existing solvers.
We also design a new non-minimal method, running on an arbitrary number of lines, to boost the performance in local optimization.
arXiv Detail & Related papers (2023-08-21T13:03:25Z) - Proximal Point Imitation Learning [48.50107891696562]
We develop new algorithms with rigorous efficiency guarantees for infinite horizon imitation learning.
We leverage classical tools from optimization, in particular, the proximal-point method (PPM) and dual smoothing.
We achieve convincing empirical performance for both linear and neural network function approximation.
arXiv Detail & Related papers (2022-09-22T12:40:21Z) - Bayesian Optimization under Stochastic Delayed Feedback [36.16843889404038]
Existing BO methods assume that the function evaluation (feedback) is available to the learner immediately or after a fixed delay.
We propose algorithms with sub-linear regret guarantees that efficiently address the dilemma of selecting new function queries while waiting for delayed feedback.
arXiv Detail & Related papers (2022-06-19T07:34:08Z) - High-probability Bounds for Non-Convex Stochastic Optimization with
Heavy Tails [55.561406656549686]
We consider non- Hilbert optimization using first-order algorithms for which the gradient estimates may have tails.
We show that a combination of gradient, momentum, and normalized gradient descent convergence to critical points in high-probability with best-known iteration for smooth losses.
arXiv Detail & Related papers (2021-06-28T00:17:01Z) - A New One-Point Residual-Feedback Oracle For Black-Box Learning and
Control [28.679167097106813]
We propose a novel one-point feedback scheme that queries the function value once at each iteration and estimates the gradient using the residual between two consecutive points.
We show that the proposed algorithm achieves the same convergence rate as that of two-point scheme with uncontrollable data samples.
arXiv Detail & Related papers (2020-06-18T19:31:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.