Oracle Efficient Algorithms for Groupwise Regret
- URL: http://arxiv.org/abs/2310.04652v1
- Date: Sat, 7 Oct 2023 02:17:22 GMT
- Title: Oracle Efficient Algorithms for Groupwise Regret
- Authors: Krishna Acharya, Eshwar Ram Arunachaleswaran, Sampath Kannan, Aaron
Roth, Juba Ziani
- Abstract summary: We show that a simple modification of the sleeping experts technique of [Blum & Lykouris] yields an efficient reduction to the well-understood problem of diminishing external regret absent group considerations.
We find that uniformly across groups, our algorithm gives substantial error improvements compared to running a standard online linear regression algorithm with no groupwise regret guarantees.
- Score: 7.840453701379554
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of online prediction, in which at each time step $t$, an
individual $x_t$ arrives, whose label we must predict. Each individual is
associated with various groups, defined based on their features such as age,
sex, race etc., which may intersect. Our goal is to make predictions that have
regret guarantees not just overall but also simultaneously on each sub-sequence
comprised of the members of any single group. Previous work such as [Blum &
Lykouris] and [Lee et al] provide attractive regret guarantees for these
problems; however, these are computationally intractable on large model
classes. We show that a simple modification of the sleeping experts technique
of [Blum & Lykouris] yields an efficient reduction to the well-understood
problem of obtaining diminishing external regret absent group considerations.
Our approach gives similar regret guarantees compared to [Blum & Lykouris];
however, we run in time linear in the number of groups, and are
oracle-efficient in the hypothesis class. This in particular implies that our
algorithm is efficient whenever the number of groups is polynomially bounded
and the external-regret problem can be solved efficiently, an improvement on
[Blum & Lykouris]'s stronger condition that the model class must be small. Our
approach can handle online linear regression and online combinatorial
optimization problems like online shortest paths. Beyond providing theoretical
regret bounds, we evaluate this algorithm with an extensive set of experiments
on synthetic data and on two real data sets -- Medical costs and the Adult
income dataset, both instantiated with intersecting groups defined in terms of
race, sex, and other demographic characteristics. We find that uniformly across
groups, our algorithm gives substantial error improvements compared to running
a standard online linear regression algorithm with no groupwise regret
guarantees.
Related papers
- Group-wise oracle-efficient algorithms for online multi-group learning [12.664869982542895]
We study the problem of online multi-group learning, a learning model in which an online learner must simultaneously achieve small prediction regret on a large collection of subsequences corresponding to a family of groups.
In this paper, we design such oracle-efficient algorithms with sublinear regret under a variety of settings, including adversarial and adversarial transductive settings.
arXiv Detail & Related papers (2024-06-07T23:00:02Z) - Cooperative Thresholded Lasso for Sparse Linear Bandit [6.52540785559241]
We present a novel approach to address the multi-agent sparse contextual linear bandit problem.
It is first algorithm that tackles row-wise distributed data in sparse linear bandits.
It is widely applicable to high-dimensional multi-agent problems where efficient feature extraction is critical for minimizing regret.
arXiv Detail & Related papers (2023-05-30T16:05:44Z) - Multicalibrated Regression for Downstream Fairness [17.084765209458762]
We show how to take a regression function $hatf$ that is appropriately multicalibrated'' and efficiently post-process it.
The post-processing requires no labeled data, and only a modest amount of unlabeled data and computation.
arXiv Detail & Related papers (2022-09-15T14:16:01Z) - Efficient and Near-Optimal Smoothed Online Learning for Generalized
Linear Functions [28.30744223973527]
We give a computationally efficient algorithm that is the first to enjoy the statistically optimal log(T/sigma) regret for realizable K-wise linear classification.
We develop a novel characterization of the geometry of the disagreement region induced by generalized linear classifiers.
arXiv Detail & Related papers (2022-05-25T21:31:36Z) - Smoothed Online Learning is as Easy as Statistical Learning [77.00766067963195]
We provide the first oracle-efficient, no-regret algorithms in this setting.
We show that if a function class is learnable in the classical setting, then there is an oracle-efficient, no-regret algorithm for contextual bandits.
arXiv Detail & Related papers (2022-02-09T19:22:34Z) - Optimal Clustering with Bandit Feedback [57.672609011609886]
This paper considers the problem of online clustering with bandit feedback.
It includes a novel stopping rule for sequential testing that circumvents the need to solve any NP-hard weighted clustering problem as its subroutines.
We show through extensive simulations on synthetic and real-world datasets that BOC's performance matches the lower boundally, and significantly outperforms a non-adaptive baseline algorithm.
arXiv Detail & Related papers (2022-02-09T06:05:05Z) - Efficient and Optimal Algorithms for Contextual Dueling Bandits under
Realizability [59.81339109121384]
We study the $K$ contextual dueling bandit problem, a sequential decision making setting in which the learner uses contextual information to make two decisions, but only observes emphpreference-based feedback suggesting that one decision was better than the other.
We provide a new algorithm that achieves the optimal regret rate for a new notion of best response regret, which is a strictly stronger performance measure than those considered in prior works.
arXiv Detail & Related papers (2021-11-24T07:14:57Z) - Efficient First-Order Contextual Bandits: Prediction, Allocation, and
Triangular Discrimination [82.52105963476703]
A recurring theme in statistical learning, online learning, and beyond is that faster convergence rates are possible for problems with low noise.
First-order guarantees are relatively well understood in statistical and online learning.
We show that the logarithmic loss and an information-theoretic quantity called the triangular discrimination play a fundamental role in obtaining first-order guarantees.
arXiv Detail & Related papers (2021-07-05T19:20:34Z) - Beyond UCB: Optimal and Efficient Contextual Bandits with Regression
Oracles [112.89548995091182]
We provide the first universal and optimal reduction from contextual bandits to online regression.
Our algorithm requires no distributional assumptions beyond realizability, and works even when contexts are chosen adversarially.
arXiv Detail & Related papers (2020-02-12T11:33:46Z) - Optimal Clustering from Noisy Binary Feedback [75.17453757892152]
We study the problem of clustering a set of items from binary user feedback.
We devise an algorithm with a minimal cluster recovery error rate.
For adaptive selection, we develop an algorithm inspired by the derivation of the information-theoretical error lower bounds.
arXiv Detail & Related papers (2019-10-14T09:18:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.