Better Boosting with Bandits for Online Learning
- URL: http://arxiv.org/abs/2001.06105v1
- Date: Thu, 16 Jan 2020 22:48:22 GMT
- Title: Better Boosting with Bandits for Online Learning
- Authors: Nikolaos Nikolaou, Joseph Mellor, Nikunj C. Oza, Gavin Brown
- Abstract summary: In learning, calibration is achieved by reserving part of the training data for training the calibrator function.
In the online setting a decision needs to be made on each round: shall the new example(s) be used to update the parameters of the ensemble or those of the calibrator.
We demonstrate superior performance to uncalibrated and naively-calibrated on-line boosting ensembles in terms of probability estimation.
- Score: 0.5543867614999908
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Probability estimates generated by boosting ensembles are poorly calibrated
because of the margin maximization nature of the algorithm. The outputs of the
ensemble need to be properly calibrated before they can be used as probability
estimates. In this work, we demonstrate that online boosting is also prone to
producing distorted probability estimates. In batch learning, calibration is
achieved by reserving part of the training data for training the calibrator
function. In the online setting, a decision needs to be made on each round:
shall the new example(s) be used to update the parameters of the ensemble or
those of the calibrator. We proceed to resolve this decision with the aid of
bandit optimization algorithms. We demonstrate superior performance to
uncalibrated and naively-calibrated on-line boosting ensembles in terms of
probability estimation. Our proposed mechanism can be easily adapted to other
tasks(e.g. cost-sensitive classification) and is robust to the choice of
hyperparameters of both the calibrator and the ensemble.
Related papers
- Optimizing Estimators of Squared Calibration Errors in Classification [2.3020018305241337]
We propose a mean-squared error-based risk that enables the comparison and optimization of estimators of squared calibration errors.
Our approach advocates for a training-validation-testing pipeline when estimating a calibration error.
arXiv Detail & Related papers (2024-10-09T15:58:06Z) - Calibration-Disentangled Learning and Relevance-Prioritized Reranking for Calibrated Sequential Recommendation [18.913912876509187]
Calibrated recommendation aims to maintain personalized proportions of categories within recommendations.
Previous methods typically leverage reranking algorithms to calibrate recommendations after training a model.
We propose LeapRec, a novel approach for the calibrated sequential recommendation.
arXiv Detail & Related papers (2024-08-04T22:23:09Z) - Towards Certification of Uncertainty Calibration under Adversarial Attacks [96.48317453951418]
We show that attacks can significantly harm calibration, and thus propose certified calibration as worst-case bounds on calibration under adversarial perturbations.
We propose novel calibration attacks and demonstrate how they can improve model calibration through textitadversarial calibration training
arXiv Detail & Related papers (2024-05-22T18:52:09Z) - Calibration by Distribution Matching: Trainable Kernel Calibration
Metrics [56.629245030893685]
We introduce kernel-based calibration metrics that unify and generalize popular forms of calibration for both classification and regression.
These metrics admit differentiable sample estimates, making it easy to incorporate a calibration objective into empirical risk minimization.
We provide intuitive mechanisms to tailor calibration metrics to a decision task, and enforce accurate loss estimation and no regret decisions.
arXiv Detail & Related papers (2023-10-31T06:19:40Z) - Hidden Heterogeneity: When to Choose Similarity-Based Calibration [12.788224825185633]
Black-box calibration methods are unable to detect subpopulations where calibration could improve prediction accuracy.
The paper proposes a quantitative measure for hidden heterogeneity (HH)
Experiments show that the improvements in calibration achieved by similarity-based calibration methods correlate with the amount of HH present and, given sufficient calibration data, generally exceed calibration achieved by global methods.
arXiv Detail & Related papers (2022-02-03T20:43:25Z) - Online Calibrated and Conformal Prediction Improves Bayesian Optimization [10.470326550507117]
This paper studies which uncertainties are needed in model-based decision-making and in Bayesian optimization.
Maintaining calibration, however, can be challenging when the data is non-stationary and depends on our actions.
We propose using simple algorithms based on online learning to provably maintain calibration on non-i.i.d. data.
arXiv Detail & Related papers (2021-12-08T23:26:23Z) - Calibrating Predictions to Decisions: A Novel Approach to Multi-Class
Calibration [118.26862029820447]
We introduce a new notion -- emphdecision calibration -- that requires the predicted distribution and true distribution to be indistinguishable'' to a set of downstream decision-makers.
Decision calibration improves decision-making on skin lesions and ImageNet classification with modern neural network.
arXiv Detail & Related papers (2021-07-12T20:17:28Z) - Localized Calibration: Metrics and Recalibration [133.07044916594361]
We propose a fine-grained calibration metric that spans the gap between fully global and fully individualized calibration.
We then introduce a localized recalibration method, LoRe, that improves the LCE better than existing recalibration methods.
arXiv Detail & Related papers (2021-02-22T07:22:12Z) - Unsupervised Calibration under Covariate Shift [92.02278658443166]
We introduce the problem of calibration under domain shift and propose an importance sampling based approach to address it.
We evaluate and discuss the efficacy of our method on both real-world datasets and synthetic datasets.
arXiv Detail & Related papers (2020-06-29T21:50:07Z) - Calibration of Neural Networks using Splines [51.42640515410253]
Measuring calibration error amounts to comparing two empirical distributions.
We introduce a binning-free calibration measure inspired by the classical Kolmogorov-Smirnov (KS) statistical test.
Our method consistently outperforms existing methods on KS error as well as other commonly used calibration measures.
arXiv Detail & Related papers (2020-06-23T07:18:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.