Optimal Tracking in Prediction with Expert Advice
- URL: http://arxiv.org/abs/2208.03708v1
- Date: Sun, 7 Aug 2022 12:29:54 GMT
- Title: Optimal Tracking in Prediction with Expert Advice
- Authors: Hakan Gokcesu, Suleyman S. Kozat
- Abstract summary: We study the prediction with expert advice setting, where the aim is to produce a decision by combining the decisions generated by a set of experts.
We achieve the min-max optimal dynamic regret under the prediction with expert advice setting.
Our algorithms are the first to produce such universally optimal, adaptive and truly online guarantees with no prior knowledge.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the prediction with expert advice setting, where the aim is to
produce a decision by combining the decisions generated by a set of experts,
e.g., independently running algorithms. We achieve the min-max optimal dynamic
regret under the prediction with expert advice setting, i.e., we can compete
against time-varying (not necessarily fixed) combinations of expert decisions
in an optimal manner. Our end-algorithm is truly online with no prior
information, such as the time horizon or loss ranges, which are commonly used
by different algorithms in the literature. Both our regret guarantees and the
min-max lower bounds are derived with the general consideration that the expert
losses can have time-varying properties and are possibly unbounded. Our
algorithm can be adapted for restrictive scenarios regarding both loss feedback
and decision making. Our guarantees are universal, i.e., our end-algorithm can
provide regret guarantee against any competitor sequence in a min-max optimal
manner with logarithmic complexity. Note that, to our knowledge, for the
prediction with expert advice problem, our algorithms are the first to produce
such universally optimal, adaptive and truly online guarantees with no prior
knowledge.
Related papers
- Active Ranking of Experts Based on their Performances in Many Tasks [72.96112117037465]
We consider the problem of ranking n experts based on their performances on d tasks.
We make a monotonicity assumption stating that for each pair of experts, one outperforms the other on all tasks.
arXiv Detail & Related papers (2023-06-05T06:55:39Z) - Streaming Algorithms for Learning with Experts: Deterministic Versus
Robust [62.98860182111096]
In the online learning with experts problem, an algorithm must make a prediction about an outcome on each of $T$ days (or times)
The goal is to make a prediction with the minimum cost, specifically compared to the best expert in the set.
We show a space lower bound of $widetildeOmegaleft(fracnMRTright)$ for any deterministic algorithm that achieves regret $R$ when the best expert makes $M$ mistakes.
arXiv Detail & Related papers (2023-03-03T04:39:53Z) - Memory Bounds for the Experts Problem [53.67419690563877]
Online learning with expert advice is a fundamental problem of sequential prediction.
The goal is to process predictions, and make a prediction with the minimum cost.
An algorithm is judged by how well it does compared to the best expert in the set.
arXiv Detail & Related papers (2022-04-21T01:22:18Z) - Second Order Regret Bounds Against Generalized Expert Sequences under
Partial Bandit Feedback [0.0]
We study the problem of expert advice under partial bandit feedback setting and create a sequential minimax optimal algorithm.
Our algorithm works with a more general partial monitoring setting, where, in contrast to the classical bandit feedback, the losses can be revealed in an adversarial manner.
arXiv Detail & Related papers (2022-04-13T22:48:12Z) - Double Coverage with Machine-Learned Advice [100.23487145400833]
We study the fundamental online $k$-server problem in a learning-augmented setting.
We show that our algorithm achieves for any k an almost optimal consistency-robustness tradeoff.
arXiv Detail & Related papers (2021-03-02T11:04:33Z) - Optimal Robustness-Consistency Trade-offs for Learning-Augmented Online
Algorithms [85.97516436641533]
We study the problem of improving the performance of online algorithms by incorporating machine-learned predictions.
The goal is to design algorithms that are both consistent and robust.
We provide the first set of non-trivial lower bounds for competitive analysis using machine-learned predictions.
arXiv Detail & Related papers (2020-10-22T04:51:01Z) - A Generalized Online Algorithm for Translation and Scale Invariant
Prediction with Expert Advice [0.0]
We study the expected regret of our algorithm against a generic competition class in the sequential prediction by expert advice problem.
Our regret bounds are stable under arbitrary scalings and translations of the losses.
arXiv Detail & Related papers (2020-09-09T15:45:28Z) - Optimal anytime regret with two experts [11.959180302329358]
We design the first minimax optimal algorithm for minimizing regret in the anytime setting.
The algorithm is designed by considering a continuous analogue of the regret problem, which is solved using ideas from calculus.
arXiv Detail & Related papers (2020-02-20T20:04:32Z) - Toward Optimal Adversarial Policies in the Multiplicative Learning
System with a Malicious Expert [87.12201611818698]
We consider a learning system that combines experts' advice to predict a sequence of true outcomes.
It is assumed that one of the experts is malicious and aims to impose the maximum loss on the system.
We show that a simple greedy policy of always reporting false prediction is optimal with an approximation ratio of $1+O(sqrtfracln NN)$.
For the online setting where the malicious expert can adaptively make its decisions, we show that the optimal online policy can be efficiently computed by solving a dynamic program in $O(N3)$.
arXiv Detail & Related papers (2020-01-02T18:04:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.