Memory Bounds for the Experts Problem
- URL: http://arxiv.org/abs/2204.09837v1
- Date: Thu, 21 Apr 2022 01:22:18 GMT
- Title: Memory Bounds for the Experts Problem
- Authors: Vaidehi Srinivas, David P. Woodruff, Ziyu Xu, Samson Zhou
- Abstract summary: Online learning with expert advice is a fundamental problem of sequential prediction.
The goal is to process predictions, and make a prediction with the minimum cost.
An algorithm is judged by how well it does compared to the best expert in the set.
- Score: 53.67419690563877
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Online learning with expert advice is a fundamental problem of sequential
prediction. In this problem, the algorithm has access to a set of $n$ "experts"
who make predictions on each day. The goal on each day is to process these
predictions, and make a prediction with the minimum cost. After making a
prediction, the algorithm sees the actual outcome on that day, updates its
state, and then moves on to the next day. An algorithm is judged by how well it
does compared to the best expert in the set.
The classical algorithm for this problem is the multiplicative weights
algorithm. However, every application, to our knowledge, relies on storing
weights for every expert, and uses $\Omega(n)$ memory. There is little work on
understanding the memory required to solve the online learning with expert
advice problem, or run standard sequential prediction algorithms, in natural
streaming models, which is especially important when the number of experts, as
well as the number of days on which the experts make predictions, is large.
We initiate the study of the learning with expert advice problem in the
streaming setting, and show lower and upper bounds. Our lower bound for i.i.d.,
random order, and adversarial order streams uses a reduction to a custom-built
problem using a novel masking technique, to show a smooth trade-off for regret
versus memory. Our upper bounds show novel ways to run standard sequential
prediction algorithms in rounds on small "pools" of experts, thus reducing the
necessary memory. For random-order streams, we show that our upper bound is
tight up to low order terms. We hope that these results and techniques will
have broad applications in online learning, and can inspire algorithms based on
standard sequential prediction techniques, like multiplicative weights, for a
wide range of other problems in the memory-constrained setting.
Related papers
- Learning-Augmented Algorithms with Explicit Predictors [67.02156211760415]
Recent advances in algorithmic design show how to utilize predictions obtained by machine learning models from past and present data.
Prior research in this context was focused on a paradigm where the predictor is pre-trained on past data and then used as a black box.
In this work, we unpack the predictor and integrate the learning problem it gives rise for within the algorithmic challenge.
arXiv Detail & Related papers (2024-03-12T08:40:21Z) - Improved Frequency Estimation Algorithms with and without Predictions [22.382900492405938]
Estimating frequencies of elements appearing in a data stream is a key task in large-scale data analysis.
We give a novel algorithm, which theoretically outperforms the learning based algorithm of Hsu et al. without the use of any predictions.
arXiv Detail & Related papers (2023-12-12T18:59:06Z) - Streaming Algorithms for Learning with Experts: Deterministic Versus
Robust [62.98860182111096]
In the online learning with experts problem, an algorithm must make a prediction about an outcome on each of $T$ days (or times)
The goal is to make a prediction with the minimum cost, specifically compared to the best expert in the set.
We show a space lower bound of $widetildeOmegaleft(fracnMRTright)$ for any deterministic algorithm that achieves regret $R$ when the best expert makes $M$ mistakes.
arXiv Detail & Related papers (2023-03-03T04:39:53Z) - Algorithms with Prediction Portfolios [23.703372221079306]
We study the use of multiple predictors for a number of fundamental problems, including matching, load balancing, and non-clairvoyant scheduling.
For each of these problems we introduce new algorithms that take advantage of multiple predictors, and prove bounds on the resulting performance.
arXiv Detail & Related papers (2022-10-22T12:58:07Z) - Robustification of Online Graph Exploration Methods [59.50307752165016]
We study a learning-augmented variant of the classical, notoriously hard online graph exploration problem.
We propose an algorithm that naturally integrates predictions into the well-known Nearest Neighbor (NN) algorithm.
arXiv Detail & Related papers (2021-12-10T10:02:31Z) - Double Coverage with Machine-Learned Advice [100.23487145400833]
We study the fundamental online $k$-server problem in a learning-augmented setting.
We show that our algorithm achieves for any k an almost optimal consistency-robustness tradeoff.
arXiv Detail & Related papers (2021-03-02T11:04:33Z) - Online Paging with a Vanishing Regret [6.520803851931362]
This paper considers a variant of the online paging problem, where the online algorithm has access to multiple predictors, each producing a sequence of predictions for the page arrival times.
The predictors may have occasional prediction errors and it is assumed that at least one of them makes a sublinear number of prediction errors in total.
Our main result states that this assumption suffices for the design of a randomized online algorithm whose time-average regret with respect to the optimal offline algorithm tends to zero as the time tends to infinity.
arXiv Detail & Related papers (2020-11-18T18:17:49Z) - Malicious Experts versus the multiplicative weights algorithm in online
prediction [85.62472761361107]
We consider a prediction problem with two experts and a forecaster.
We assume that one of the experts is honest and makes correct prediction with probability $mu$ at each round.
The other one is malicious, who knows true outcomes at each round and makes predictions in order to maximize the loss of the forecaster.
arXiv Detail & Related papers (2020-03-18T20:12:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.