Loss Minimization through the Lens of Outcome Indistinguishability
- URL: http://arxiv.org/abs/2210.08649v1
- Date: Sun, 16 Oct 2022 22:25:27 GMT
- Title: Loss Minimization through the Lens of Outcome Indistinguishability
- Authors: Parikshit Gopalan, Lunjia Hu, Michael P. Kim, Omer Reingold, Udi
Wieder
- Abstract summary: We present a new perspective on convex loss and the recent notion of Omniprediction.
By design, Loss OI implies omniprediction in a direct and intuitive manner.
We show that Loss OI for the important set of losses arising from Generalized Models, without requiring full multicalibration.
- Score: 11.709566373491619
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We present a new perspective on loss minimization and the recent notion of
Omniprediction through the lens of Outcome Indistingusihability. For a
collection of losses and hypothesis class, omniprediction requires that a
predictor provide a loss-minimization guarantee simultaneously for every loss
in the collection compared to the best (loss-specific) hypothesis in the class.
We present a generic template to learn predictors satisfying a guarantee we
call Loss Outcome Indistinguishability. For a set of statistical tests--based
on a collection of losses and hypothesis class--a predictor is Loss OI if it is
indistinguishable (according to the tests) from Nature's true probabilities
over outcomes. By design, Loss OI implies omniprediction in a direct and
intuitive manner. We simplify Loss OI further, decomposing it into a
calibration condition plus multiaccuracy for a class of functions derived from
the loss and hypothesis classes. By careful analysis of this class, we give
efficient constructions of omnipredictors for interesting classes of loss
functions, including non-convex losses.
This decomposition highlights the utility of a new multi-group fairness
notion that we call calibrated multiaccuracy, which lies in between
multiaccuracy and multicalibration. We show that calibrated multiaccuracy
implies Loss OI for the important set of convex losses arising from Generalized
Linear Models, without requiring full multicalibration. For such losses, we
show an equivalence between our computational notion of Loss OI and a geometric
notion of indistinguishability, formulated as Pythagorean theorems in the
associated Bregman divergence. We give an efficient algorithm for calibrated
multiaccuracy with computational complexity comparable to that of
multiaccuracy. In all, calibrated multiaccuracy offers an interesting tradeoff
point between efficiency and generality in the omniprediction landscape.
Related papers
- High-dimensional logistic regression with missing data: Imputation, regularization, and universality [7.167672851569787]
We study high-dimensional, ridge-regularized logistic regression.
We provide exact characterizations of both the prediction error and the estimation error.
arXiv Detail & Related papers (2024-10-01T21:41:21Z) - EnsLoss: Stochastic Calibrated Loss Ensembles for Preventing Overfitting in Classification [1.3778851745408134]
We propose a novel ensemble method, namely EnsLoss, to combine loss functions within the Empirical risk minimization framework.
We first transform the CC conditions of losses into loss-derivatives, thereby bypassing the need for explicit loss functions.
We theoretically establish the statistical consistency of our approach and provide insights into its benefits.
arXiv Detail & Related papers (2024-09-02T02:40:42Z) - Orthogonal Causal Calibration [55.28164682911196]
We prove generic upper bounds on the calibration error of any causal parameter estimate $theta$ with respect to any loss $ell$.
We use our bound to analyze the convergence of two sample splitting algorithms for causal calibration.
arXiv Detail & Related papers (2024-06-04T03:35:25Z) - Cross-Entropy Loss Functions: Theoretical Analysis and Applications [27.3569897539488]
We present a theoretical analysis of a broad family of loss functions, that includes cross-entropy (or logistic loss), generalized cross-entropy, the mean absolute error and other cross-entropy-like loss functions.
We show that these loss functions are beneficial in the adversarial setting by proving that they admit $H$-consistency bounds.
This leads to new adversarial robustness algorithms that consist of minimizing a regularized smooth adversarial comp-sum loss.
arXiv Detail & Related papers (2023-04-14T17:58:23Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - Beyond calibration: estimating the grouping loss of modern neural
networks [68.8204255655161]
Proper scoring rule theory shows that given the calibration loss, the missing piece to characterize individual errors is the grouping loss.
We show that modern neural network architectures in vision and NLP exhibit grouping loss, notably in distribution shifts settings.
arXiv Detail & Related papers (2022-10-28T07:04:20Z) - Omnipredictors for Constrained Optimization [5.969079530101132]
We show how to obtain omnipredictors for constrained optimization problems, relying on appropriate variants of multicalibration.
We also investigate the implications of this notion when the constraints used are so-called group fairness notions.
arXiv Detail & Related papers (2022-09-15T17:04:49Z) - Efficient and Differentiable Conformal Prediction with General Function
Classes [96.74055810115456]
We propose a generalization of conformal prediction to multiple learnable parameters.
We show that it achieves approximate valid population coverage and near-optimal efficiency within class.
Experiments show that our algorithm is able to learn valid prediction sets and improve the efficiency significantly.
arXiv Detail & Related papers (2022-02-22T18:37:23Z) - Omnipredictors [19.735769148626588]
Loss minimization is a dominant paradigm in machine learning.
We introduce the notion of an ($mathcalL,mathcalC$)-omnipredictor, which could be used to optimize any loss in a family.
We show that such "loss-oblivious'' learning is feasible through a connection to multicalibration.
arXiv Detail & Related papers (2021-09-11T23:28:49Z) - A Symmetric Loss Perspective of Reliable Machine Learning [87.68601212686086]
We review how a symmetric loss can yield robust classification from corrupted labels in balanced error rate (BER) minimization.
We demonstrate how the robust AUC method can benefit natural language processing in the problem where we want to learn only from relevant keywords.
arXiv Detail & Related papers (2021-01-05T06:25:47Z) - Calibrated Surrogate Losses for Adversarially Robust Classification [92.37268323142307]
We show that no convex surrogate loss is respect with respect to adversarial 0-1 loss when restricted to linear models.
We also show that if the underlying distribution satisfies the Massart's noise condition, convex losses can also be calibrated in the adversarial setting.
arXiv Detail & Related papers (2020-05-28T02:40:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.