A Generalized Online Algorithm for Translation and Scale Invariant
Prediction with Expert Advice
- URL: http://arxiv.org/abs/2009.04372v1
- Date: Wed, 9 Sep 2020 15:45:28 GMT
- Title: A Generalized Online Algorithm for Translation and Scale Invariant
Prediction with Expert Advice
- Authors: Kaan Gokcesu, Hakan Gokcesu
- Abstract summary: We study the expected regret of our algorithm against a generic competition class in the sequential prediction by expert advice problem.
Our regret bounds are stable under arbitrary scalings and translations of the losses.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we aim to create a completely online algorithmic framework for
prediction with expert advice that is translation-free and scale-free of the
expert losses. Our goal is to create a generalized algorithm that is suitable
for use in a wide variety of applications. For this purpose, we study the
expected regret of our algorithm against a generic competition class in the
sequential prediction by expert advice problem, where the expected regret
measures the difference between the losses of our prediction algorithm and the
losses of the 'best' expert selection strategy in the competition. We design
our algorithm using the universal prediction perspective to compete against a
specified class of expert selection strategies, which is not necessarily a
fixed expert selection. The class of expert selection strategies that we want
to compete against is purely determined by the specific application at hand and
is left generic, which makes our generalized algorithm suitable for use in many
different problems. We show that no preliminary knowledge about the loss
sequence is required by our algorithm and its performance bounds, which are
second order, expressed in terms of sums of squared losses. Our regret bounds
are stable under arbitrary scalings and translations of the losses.
Related papers
- Data Dependent Regret Guarantees Against General Comparators for Full or
Bandit Feedback [0.0]
We study the adversarial online learning problem and create a completely online algorithmic framework that has data dependent regret guarantees.
Our algorithm works from a universal prediction perspective and the performance measure used is the expected regret against arbitrary comparator sequences.
arXiv Detail & Related papers (2023-03-12T00:18:46Z) - Streaming Algorithms for Learning with Experts: Deterministic Versus
Robust [62.98860182111096]
In the online learning with experts problem, an algorithm must make a prediction about an outcome on each of $T$ days (or times)
The goal is to make a prediction with the minimum cost, specifically compared to the best expert in the set.
We show a space lower bound of $widetildeOmegaleft(fracnMRTright)$ for any deterministic algorithm that achieves regret $R$ when the best expert makes $M$ mistakes.
arXiv Detail & Related papers (2023-03-03T04:39:53Z) - Minimalistic Predictions to Schedule Jobs with Online Precedence
Constraints [117.8317521974783]
We consider non-clairvoyant scheduling with online precedence constraints.
An algorithm is oblivious to any job dependencies and learns about a job only if all of its predecessors have been completed.
arXiv Detail & Related papers (2023-01-30T13:17:15Z) - Optimal Tracking in Prediction with Expert Advice [0.0]
We study the prediction with expert advice setting, where the aim is to produce a decision by combining the decisions generated by a set of experts.
We achieve the min-max optimal dynamic regret under the prediction with expert advice setting.
Our algorithms are the first to produce such universally optimal, adaptive and truly online guarantees with no prior knowledge.
arXiv Detail & Related papers (2022-08-07T12:29:54Z) - Second Order Regret Bounds Against Generalized Expert Sequences under
Partial Bandit Feedback [0.0]
We study the problem of expert advice under partial bandit feedback setting and create a sequential minimax optimal algorithm.
Our algorithm works with a more general partial monitoring setting, where, in contrast to the classical bandit feedback, the losses can be revealed in an adversarial manner.
arXiv Detail & Related papers (2022-04-13T22:48:12Z) - Robustification of Online Graph Exploration Methods [59.50307752165016]
We study a learning-augmented variant of the classical, notoriously hard online graph exploration problem.
We propose an algorithm that naturally integrates predictions into the well-known Nearest Neighbor (NN) algorithm.
arXiv Detail & Related papers (2021-12-10T10:02:31Z) - Generalized Translation and Scale Invariant Online Algorithm for
Adversarial Multi-Armed Bandits [0.0]
We study the adversarial multi-armed bandit problem and create a completely online algorithmic framework that is invariant under arbitrary translations and scales of the arm losses.
Our algorithm works from a universal prediction perspective and the performance measure used is the expected regret against arbitrary arm selection sequences.
arXiv Detail & Related papers (2021-09-19T20:13:59Z) - Bayesian decision-making under misspecified priors with applications to
meta-learning [64.38020203019013]
Thompson sampling and other sequential decision-making algorithms are popular approaches to tackle explore/exploit trade-offs in contextual bandits.
We show that performance degrades gracefully with misspecified priors.
arXiv Detail & Related papers (2021-07-03T23:17:26Z) - Double Coverage with Machine-Learned Advice [100.23487145400833]
We study the fundamental online $k$-server problem in a learning-augmented setting.
We show that our algorithm achieves for any k an almost optimal consistency-robustness tradeoff.
arXiv Detail & Related papers (2021-03-02T11:04:33Z) - Optimal Robustness-Consistency Trade-offs for Learning-Augmented Online
Algorithms [85.97516436641533]
We study the problem of improving the performance of online algorithms by incorporating machine-learned predictions.
The goal is to design algorithms that are both consistent and robust.
We provide the first set of non-trivial lower bounds for competitive analysis using machine-learned predictions.
arXiv Detail & Related papers (2020-10-22T04:51:01Z) - No-Regret and Incentive-Compatible Online Learning [29.267666165169324]
We study online learning settings in which experts act strategically to maximize their influence on the learning algorithm's predictions.
We want the learning algorithm to be no-regret with respect to the best fixed expert in hindsight.
We provide algorithms that achieve no regret and incentive compatibility for experts for both the full and partial information settings.
arXiv Detail & Related papers (2020-02-20T16:21:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.