Agnostic learning with unknown utilities
- URL: http://arxiv.org/abs/2104.08482v1
- Date: Sat, 17 Apr 2021 08:22:04 GMT
- Title: Agnostic learning with unknown utilities
- Authors: Kush Bhatia, Peter L. Bartlett, Anca D. Dragan, Jacob Steinhardt
- Abstract summary: In many real-world problems, the utility of a decision depends on the underlying context $x$ and decision $y$.
We study this as agnostic learning with unknown utilities.
We show that estimating the utilities of only the sampled points$S$ suffices to learn a decision function which generalizes well.
- Score: 70.14742836006042
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traditional learning approaches for classification implicitly assume that
each mistake has the same cost. In many real-world problems though, the utility
of a decision depends on the underlying context $x$ and decision $y$. However,
directly incorporating these utilities into the learning objective is often
infeasible since these can be quite complex and difficult for humans to
specify.
We formally study this as agnostic learning with unknown utilities: given a
dataset $S = \{x_1, \ldots, x_n\}$ where each data point $x_i \sim
\mathcal{D}$, the objective of the learner is to output a function $f$ in some
class of decision functions $\mathcal{F}$ with small excess risk. This risk
measures the performance of the output predictor $f$ with respect to the best
predictor in the class $\mathcal{F}$ on the unknown underlying utility $u^*$.
This utility $u^*$ is not assumed to have any specific structure. This raises
an interesting question whether learning is even possible in our setup, given
that obtaining a generalizable estimate of utility $u^*$ might not be possible
from finitely many samples. Surprisingly, we show that estimating the utilities
of only the sampled points~$S$ suffices to learn a decision function which
generalizes well.
We study mechanisms for eliciting information which allow a learner to
estimate the utilities $u^*$ on the set $S$. We introduce a family of
elicitation mechanisms by generalizing comparisons, called the $k$-comparison
oracle, which enables the learner to ask for comparisons across $k$ different
inputs $x$ at once. We show that the excess risk in our agnostic learning
framework decreases at a rate of $O\left(\frac{1}{k} \right)$. This result
brings out an interesting accuracy-elicitation trade-off -- as the order $k$ of
the oracle increases, the comparative queries become harder to elicit from
humans but allow for more accurate learning.
Related papers
- Guarantees for Nonlinear Representation Learning: Non-identical Covariates, Dependent Data, Fewer Samples [24.45016514352055]
We study the sample-complexity of learning $T+1$ functions $f_star(t) circ g_star$ from a function class $mathcal F times mathcal G$.
We show that as the number of tasks $T$ increases, both the sample requirement and risk bound converge to that of $r$-dimensional regression.
arXiv Detail & Related papers (2024-10-15T03:20:19Z) - Revisiting Agnostic PAC Learning [30.67561230812141]
PAC learning, dating back to Valiant'84 and Vapnik and Chervonenkis'64,'74, is a classic model for studying supervised learning.
Empirical Risk Minimization (ERM) is a natural learning algorithm, where one simply outputs the hypothesis from $mathcalH$ making the fewest mistakes on the training data.
We revisit PAC learning and first show that ERM is in fact sub-optimal if we treat the performance of the best hypothesis, denoted $tau:=Pr_mathcalD[hstar_math
arXiv Detail & Related papers (2024-07-29T08:20:49Z) - Agnostically Learning Multi-index Models with Queries [54.290489524576756]
We study the power of query access for the task of agnostic learning under the Gaussian distribution.
We show that query access gives significant runtime improvements over random examples for agnostically learning MIMs.
arXiv Detail & Related papers (2023-12-27T15:50:47Z) - On Optimal Learning Under Targeted Data Poisoning [48.907813854832206]
In this work we aim to characterize the smallest achievable error $epsilon=epsilon(eta)$ by the learner in the presence of such an adversary.
Remarkably, we show that the upper bound can be attained by a deterministic learner.
arXiv Detail & Related papers (2022-10-06T06:49:48Z) - There is no Accuracy-Interpretability Tradeoff in Reinforcement Learning
for Mazes [64.05903267230467]
Interpretability is an essential building block for trustworthiness in reinforcement learning systems.
We show that in certain cases, one can achieve policy interpretability while maintaining its optimality.
arXiv Detail & Related papers (2022-06-09T04:23:26Z) - TURF: A Two-factor, Universal, Robust, Fast Distribution Learning
Algorithm [64.13217062232874]
One of its most powerful and successful modalities approximates every distribution to an $ell$ distance essentially at most a constant times larger than its closest $t$-piece degree-$d_$.
We provide a method that estimates this number near-optimally, hence helps approach the best possible approximation.
arXiv Detail & Related papers (2022-02-15T03:49:28Z) - Mediated Uncoupled Learning: Learning Functions without Direct
Input-output Correspondences [80.95776331769899]
We consider the task of predicting $Y$ from $X$ when we have no paired data of them.
A naive approach is to predict $U$ from $X$ using $S_X$ and then $Y$ from $U$ using $S_Y$.
We propose a new method that avoids predicting $U$ but directly learns $Y = f(X)$ by training $f(X)$ with $S_X$ to predict $h(U)$.
arXiv Detail & Related papers (2021-07-16T22:13:29Z) - Remember What You Want to Forget: Algorithms for Machine Unlearning [37.656345901739506]
We study the problem of forgetting datapoints from a learnt model.
We provide an unlearning algorithm that can delete up to $O(n/d1/4)$ samples.
arXiv Detail & Related papers (2021-03-04T19:28:57Z) - Does generalization performance of $l^q$ regularization learning depend
on $q$? A negative example [19.945160684285003]
$lq$-regularization has been demonstrated to be an attractive technique in machine learning and statistical modeling.
We show that all $lq$ estimators for $0 infty$ attain similar generalization error bounds.
This finding tentatively reveals that, in some modeling contexts, the choice of $q$ might not have a strong impact in terms of the generalization capability.
arXiv Detail & Related papers (2013-07-25T00:48:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.