Adaptive calibration for binary classification
- URL: http://arxiv.org/abs/2107.01726v1
- Date: Sun, 4 Jul 2021 20:32:52 GMT
- Title: Adaptive calibration for binary classification
- Authors: Vladimir Vovk, Ivan Petej, and Alex Gammerman
- Abstract summary: This is important in applications of machine learning, where the quality of a trained predictor may drop significantly in the process of its exploitation.
Our techniques are based on recent work on conformal test martingales and older work on prediction with expert advice, namely tracking the best expert.
- Score: 0.20072624123275526
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This note proposes a way of making probability forecasting rules less
sensitive to changes in data distribution, concentrating on the simple case of
binary classification. This is important in applications of machine learning,
where the quality of a trained predictor may drop significantly in the process
of its exploitation. Our techniques are based on recent work on conformal test
martingales and older work on prediction with expert advice, namely tracking
the best expert.
Related papers
- Algorithms with Calibrated Machine Learning Predictions [9.18151868060576]
The field of algorithms with predictions incorporates machine learning advice in the design of online algorithms to improve real-world performance.
We propose calibration as a principled and practical tool to bridge this gap, demonstrating the benefits of calibrated advice through two case studies.
For ski rental, we design an algorithm that achieves optimal prediction-dependent performance and prove that, in high-variance settings, calibrated advice offers more effective guidance than alternative methods for uncertainty.
For job scheduling, we demonstrate that using a predictor leads to significant performance improvements over existing methods.
arXiv Detail & Related papers (2025-02-05T03:41:18Z) - Rethinking Early Stopping: Refine, Then Calibrate [49.966899634962374]
We show that calibration error and refinement error are not minimized simultaneously during training.
We introduce a new metric for early stopping and hyper parameter tuning that makes it possible to minimize refinement error during training.
Our method integrates seamlessly with any architecture and consistently improves performance across diverse classification tasks.
arXiv Detail & Related papers (2025-01-31T15:03:54Z) - Trustworthy Classification through Rank-Based Conformal Prediction Sets [9.559062601251464]
We propose a novel conformal prediction method that employs a rank-based score function suitable for classification models.
Our approach constructs prediction sets that achieve the desired coverage rate while managing their size.
Our contributions include a novel conformal prediction method, theoretical analysis, and empirical evaluation.
arXiv Detail & Related papers (2024-07-05T10:43:41Z) - Adapting Conformal Prediction to Distribution Shifts Without Labels [16.478151550456804]
Conformal prediction (CP) enables machine learning models to output prediction sets with guaranteed coverage rate.
Our goal is to improve the quality of CP-generated prediction sets using only unlabeled data from the test domain.
This is achieved by two new methods called ECP and EACP, that adjust the score function in CP according to the base model's uncertainty on the unlabeled test data.
arXiv Detail & Related papers (2024-06-03T15:16:02Z) - Conformal Prediction with Partially Labeled Data [3.895044919159418]
We propose a generalization of the conformal prediction procedure that can be applied to set-valued training and calibration data.
We prove the validity of the proposed method and present experimental studies in which it compares favorably to natural baselines.
arXiv Detail & Related papers (2023-06-01T23:10:15Z) - On the Trade-off of Intra-/Inter-class Diversity for Supervised
Pre-training [72.8087629914444]
We study the impact of the trade-off between the intra-class diversity (the number of samples per class) and the inter-class diversity (the number of classes) of a supervised pre-training dataset.
With the size of the pre-training dataset fixed, the best downstream performance comes with a balance on the intra-/inter-class diversity.
arXiv Detail & Related papers (2023-05-20T16:23:50Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - Adaptive Dimension Reduction and Variational Inference for Transductive
Few-Shot Classification [2.922007656878633]
We propose a new clustering method based on Variational Bayesian inference, further improved by Adaptive Dimension Reduction.
Our proposed method significantly improves accuracy in the realistic unbalanced transductive setting on various Few-Shot benchmarks.
arXiv Detail & Related papers (2022-09-18T10:29:02Z) - Improved Fine-tuning by Leveraging Pre-training Data: Theory and
Practice [52.11183787786718]
Fine-tuning a pre-trained model on the target data is widely used in many deep learning applications.
Recent studies have empirically shown that training from scratch has the final performance that is no worse than this pre-training strategy.
We propose a novel selection strategy to select a subset from pre-training data to help improve the generalization on the target task.
arXiv Detail & Related papers (2021-11-24T06:18:32Z) - The Hitchhiker's Guide to Prior-Shift Adaptation [41.4341627937948]
We propose a novel method to address a known issue of prior estimation methods based on confusion matrices.
Experiments on fine-grained image classification datasets provide insight into the best practice of prior shift estimation.
Applying the best practice to two tasks with naturally imbalanced priors, learning from web-crawled images and plant species classification increased the recognition accuracy by 1.1% and 3.4% respectively.
arXiv Detail & Related papers (2021-06-22T11:55:51Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Distribution-Free, Risk-Controlling Prediction Sets [112.9186453405701]
We show how to generate set-valued predictions from a black-box predictor that control the expected loss on future test points at a user-specified level.
Our approach provides explicit finite-sample guarantees for any dataset by using a holdout set to calibrate the size of the prediction sets.
arXiv Detail & Related papers (2021-01-07T18:59:33Z) - Provable tradeoffs in adversarially robust classification [96.48180210364893]
We develop and leverage new tools, including recent breakthroughs from probability theory on robust isoperimetry.
Our results reveal fundamental tradeoffs between standard and robust accuracy that grow when data is imbalanced.
arXiv Detail & Related papers (2020-06-09T09:58:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.