Parsimonious Learning-Augmented Caching
- URL: http://arxiv.org/abs/2202.04262v1
- Date: Wed, 9 Feb 2022 03:40:11 GMT
- Title: Parsimonious Learning-Augmented Caching
- Authors: Sungjin Im, Ravi Kumar, Aditya Petety, Manish Purohit
- Abstract summary: We introduce and study the setting in which the learning-augmented algorithm can utilize the predictions parsimoniously.
We show that one can achieve quantitatively similar results but only using a sublinear number of predictions.
- Score: 29.975391787684966
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning-augmented algorithms -- in which, traditional algorithms are
augmented with machine-learned predictions -- have emerged as a framework to go
beyond worst-case analysis. The overarching goal is to design algorithms that
perform near-optimally when the predictions are accurate yet retain certain
worst-case guarantees irrespective of the accuracy of the predictions. This
framework has been successfully applied to online problems such as caching
where the predictions can be used to alleviate uncertainties.
In this paper we introduce and study the setting in which the
learning-augmented algorithm can utilize the predictions parsimoniously. We
consider the caching problem -- which has been extensively studied in the
learning-augmented setting -- and show that one can achieve quantitatively
similar results but only using a sublinear number of predictions.
Related papers
- Learning-Augmented Algorithms with Explicit Predictors [67.02156211760415]
Recent advances in algorithmic design show how to utilize predictions obtained by machine learning models from past and present data.
Prior research in this context was focused on a paradigm where the predictor is pre-trained on past data and then used as a black box.
In this work, we unpack the predictor and integrate the learning problem it gives rise for within the algorithmic challenge.
arXiv Detail & Related papers (2024-03-12T08:40:21Z) - Algorithms with Prediction Portfolios [23.703372221079306]
We study the use of multiple predictors for a number of fundamental problems, including matching, load balancing, and non-clairvoyant scheduling.
For each of these problems we introduce new algorithms that take advantage of multiple predictors, and prove bounds on the resulting performance.
arXiv Detail & Related papers (2022-10-22T12:58:07Z) - Efficient and Differentiable Conformal Prediction with General Function
Classes [96.74055810115456]
We propose a generalization of conformal prediction to multiple learnable parameters.
We show that it achieves approximate valid population coverage and near-optimal efficiency within class.
Experiments show that our algorithm is able to learn valid prediction sets and improve the efficiency significantly.
arXiv Detail & Related papers (2022-02-22T18:37:23Z) - Non-Clairvoyant Scheduling with Predictions Revisited [77.86290991564829]
In non-clairvoyant scheduling, the task is to find an online strategy for scheduling jobs with a priori unknown processing requirements.
We revisit this well-studied problem in a recently popular learning-augmented setting that integrates (untrusted) predictions in algorithm design.
We show that these predictions have desired properties, admit a natural error measure as well as algorithms with strong performance guarantees.
arXiv Detail & Related papers (2022-02-21T13:18:11Z) - Learning Predictions for Algorithms with Predictions [49.341241064279714]
We introduce a general design approach for algorithms that learn predictors.
We apply techniques from online learning to learn against adversarial instances, tune robustness-consistency trade-offs, and obtain new statistical guarantees.
We demonstrate the effectiveness of our approach at deriving learning algorithms by analyzing methods for bipartite matching, page migration, ski-rental, and job scheduling.
arXiv Detail & Related papers (2022-02-18T17:25:43Z) - Robustification of Online Graph Exploration Methods [59.50307752165016]
We study a learning-augmented variant of the classical, notoriously hard online graph exploration problem.
We propose an algorithm that naturally integrates predictions into the well-known Nearest Neighbor (NN) algorithm.
arXiv Detail & Related papers (2021-12-10T10:02:31Z) - Double Coverage with Machine-Learned Advice [100.23487145400833]
We study the fundamental online $k$-server problem in a learning-augmented setting.
We show that our algorithm achieves for any k an almost optimal consistency-robustness tradeoff.
arXiv Detail & Related papers (2021-03-02T11:04:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.