The Scope of Multicalibration: Characterizing Multicalibration via
Property Elicitation
- URL: http://arxiv.org/abs/2302.08507v1
- Date: Thu, 16 Feb 2023 18:59:16 GMT
- Title: The Scope of Multicalibration: Characterizing Multicalibration via
Property Elicitation
- Authors: Georgy Noarov, Aaron Roth
- Abstract summary: We show that it is possible to produce a multicalibrated predictor for a continuous scalar distributional property $Gamma$ if and only if $Gamma$ is elicitable.
On the negative side, we show that for non-elicitable continuous properties there exist simple data distributions on which even the true distributional predictor is not calibrated.
- Score: 12.197909808303411
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We make a connection between multicalibration and property elicitation and
show that (under mild technical conditions) it is possible to produce a
multicalibrated predictor for a continuous scalar distributional property
$\Gamma$ if and only if $\Gamma$ is elicitable.
On the negative side, we show that for non-elicitable continuous properties
there exist simple data distributions on which even the true distributional
predictor is not calibrated. On the positive side, for elicitable $\Gamma$, we
give simple canonical algorithms for the batch and the online adversarial
setting, that learn a $\Gamma$-multicalibrated predictor. This generalizes past
work on multicalibrated means and quantiles, and in fact strengthens existing
online quantile multicalibration results.
To further counter-weigh our negative result, we show that if a property
$\Gamma^1$ is not elicitable by itself, but is elicitable conditionally on
another elicitable property $\Gamma^0$, then there is a canonical algorithm
that jointly multicalibrates $\Gamma^1$ and $\Gamma^0$; this generalizes past
work on mean-moment multicalibration.
Finally, as applications of our theory, we provide novel algorithmic and
impossibility results for fair (multicalibrated) risk assessment.
Related papers
- Efficient Swap Multicalibration of Elicitable Properties [41.64272548564929]
[NR23] established a surprising connection between multicalibration for an arbitrary property $Gamma$ and property elicitation.<n>In this paper, we generalize multicalibration for an elicitable property $Gamma$ from group membership functions to arbitrary bounded hypothesis classes.
arXiv Detail & Related papers (2025-11-07T01:14:39Z) - Computational-Statistical Tradeoffs at the Next-Token Prediction Barrier: Autoregressive and Imitation Learning under Misspecification [50.717692060500696]
Next-token prediction with the logarithmic loss is a cornerstone of autoregressive sequence modeling.
Next-token prediction can be made robust so as to achieve $C=tilde O(H)$, representing moderate error amplification.
No computationally efficient algorithm can achieve sub-polynomial approximation factor $C=e(log H)1-Omega(1)$.
arXiv Detail & Related papers (2025-02-18T02:52:00Z) - Variance-Dependent Regret Bounds for Non-stationary Linear Bandits [52.872628573907434]
We propose algorithms that utilize the variance of the reward distribution as well as the $B_K$, and show that they can achieve tighter regret upper bounds.
We introduce two novel algorithms: Restarted Weighted$textOFUL+$ and Restarted $textSAVE+$.
Notably, when the total variance $V_K$ is much smaller than $K$, our algorithms outperform previous state-of-the-art results on non-stationary linear bandits under different settings.
arXiv Detail & Related papers (2024-03-15T23:36:55Z) - On Computationally Efficient Multi-Class Calibration [9.032290717007065]
Project calibration gives strong guarantees for all downstream decision makers.
It ensures that the probabilities predicted by summing the probabilities assigned to labels in $T$ are close to some perfectly calibrated binary predictor.
arXiv Detail & Related papers (2024-02-12T17:25:23Z) - Distributed Extra-gradient with Optimal Complexity and Communication
Guarantees [60.571030754252824]
We consider monotone variational inequality (VI) problems in multi-GPU settings where multiple processors/workers/clients have access to local dual vectors.
Extra-gradient, which is a de facto algorithm for monotone VI problems, has not been designed to be communication-efficient.
We propose a quantized generalized extra-gradient (Q-GenX), which is an unbiased and adaptive compression method tailored to solve VIs.
arXiv Detail & Related papers (2023-08-17T21:15:04Z) - Improved Bound for Mixing Time of Parallel Tempering [4.68299658663016]
We present a new lower bound for parallel tempering on the spectral gap that has a multimodal dependence on all parameters except $log L$.
This improves the best existing bound which depends exponentially on the number of modes.
We complement our result with a hypothetical upper bound on spectral gap that has an exponential dependence on $log L$, which shows that, in some sense, our bound is tight.
arXiv Detail & Related papers (2023-04-03T19:03:22Z) - HappyMap: A Generalized Multi-calibration Method [23.086009024383024]
Multi-calibration is a powerful and evolving concept originating in the field of algorithmic fairness.
In this work, we view the term $(f(x)-y)$ as just one specific mapping, and explore the power of an enriched class of mappings.
We propose textitHappyMap, a generalization of multi-calibration, which yields a wide range of new applications.
arXiv Detail & Related papers (2023-03-08T05:05:01Z) - Distributional Hardness Against Preconditioned Lasso via Erasure-Robust
Designs [22.41443027099101]
We show that standard sparse random designs are with high probability robust to adversarial measurement erasures.
This is the first time that partial recoverability of arbitrary sparse signals under erasures has been studied in compressed sensing.
arXiv Detail & Related papers (2022-03-05T22:16:05Z) - A Law of Robustness beyond Isoperimetry [84.33752026418045]
We prove a Lipschitzness lower bound $Omega(sqrtn/p)$ of robustness of interpolating neural network parameters on arbitrary distributions.
We then show the potential benefit of overparametrization for smooth data when $n=mathrmpoly(d)$.
We disprove the potential existence of an $O(1)$-Lipschitz robust interpolating function when $n=exp(omega(d))$.
arXiv Detail & Related papers (2022-02-23T16:10:23Z) - Understanding the Under-Coverage Bias in Uncertainty Estimation [58.03725169462616]
quantile regression tends to emphunder-cover than the desired coverage level in reality.
We prove that quantile regression suffers from an inherent under-coverage bias.
Our theory reveals that this under-coverage bias stems from a certain high-dimensional parameter estimation error.
arXiv Detail & Related papers (2021-06-10T06:11:55Z) - Convergence Rates of Stochastic Gradient Descent under Infinite Noise
Variance [14.06947898164194]
Heavy tails emerge in gradient descent (SGD) in various scenarios.
We provide convergence guarantees for SGD under a state-dependent and heavy-tailed noise with a potentially infinite variance.
Our results indicate that even under heavy-tailed noise with infinite variance, SGD can converge to the global optimum.
arXiv Detail & Related papers (2021-02-20T13:45:11Z) - Almost Tight L0-norm Certified Robustness of Top-k Predictions against
Adversarial Perturbations [78.23408201652984]
Top-k predictions are used in many real-world applications such as machine learning as a service, recommender systems, and web searches.
Our work is based on randomized smoothing, which builds a provably robust classifier via randomizing an input.
For instance, our method can build a classifier that achieves a certified top-3 accuracy of 69.2% on ImageNet when an attacker can arbitrarily perturb 5 pixels of a testing image.
arXiv Detail & Related papers (2020-11-15T21:34:44Z) - Linear-Sample Learning of Low-Rank Distributions [56.59844655107251]
We show that learning $ktimes k$, rank-$r$, matrices to normalized $L_1$ distance requires $Omega(frackrepsilon2)$ samples.
We propose an algorithm that uses $cal O(frackrepsilon2log2fracepsilon)$ samples, a number linear in the high dimension, and nearly linear in the matrices, typically low, rank proofs.
arXiv Detail & Related papers (2020-09-30T19:10:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.