Calibrate and Prune: Improving Reliability of Lottery Tickets Through
Prediction Calibration
- URL: http://arxiv.org/abs/2002.03875v3
- Date: Wed, 30 Sep 2020 05:17:25 GMT
- Title: Calibrate and Prune: Improving Reliability of Lottery Tickets Through
Prediction Calibration
- Authors: Bindya Venkatesh, Jayaraman J. Thiagarajan, Kowshik Thopalli and
Prasanna Sattigeri
- Abstract summary: Supervised models with uncalibrated confidences tend to be overconfident even when making wrong prediction.
We study how explicit confidence calibration in the over- parameterized network impacts the quality of the resulting lottery tickets.
Our empirical studies reveal that including calibration mechanisms consistently lead to more effective lottery tickets.
- Score: 40.203492372949576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The hypothesis that sub-network initializations (lottery) exist within the
initializations of over-parameterized networks, which when trained in isolation
produce highly generalizable models, has led to crucial insights into network
initialization and has enabled efficient inferencing. Supervised models with
uncalibrated confidences tend to be overconfident even when making wrong
prediction. In this paper, for the first time, we study how explicit confidence
calibration in the over-parameterized network impacts the quality of the
resulting lottery tickets. More specifically, we incorporate a suite of
calibration strategies, ranging from mixup regularization, variance-weighted
confidence calibration to the newly proposed likelihood-based calibration and
normalized bin assignment strategies. Furthermore, we explore different
combinations of architectures and datasets, and make a number of key findings
about the role of confidence calibration. Our empirical studies reveal that
including calibration mechanisms consistently lead to more effective lottery
tickets, in terms of accuracy as well as empirical calibration metrics, even
when retrained using data with challenging distribution shifts with respect to
the source dataset.
Related papers
- Provably Reliable Conformal Prediction Sets in the Presence of Data Poisoning [53.42244686183879]
Conformal prediction provides model-agnostic and distribution-free uncertainty quantification.
Yet, conformal prediction is not reliable under poisoning attacks where adversaries manipulate both training and calibration data.
We propose reliable prediction sets (RPS): the first efficient method for constructing conformal prediction sets with provable reliability guarantees under poisoning.
arXiv Detail & Related papers (2024-10-13T15:37:11Z) - Towards Calibrated Deep Clustering Network [60.71776081164377]
In deep clustering, the estimated confidence for a sample belonging to a particular cluster greatly exceeds its actual prediction accuracy.
We propose a novel dual-head (calibration head and clustering head) deep clustering model that can effectively calibrate the estimated confidence and the actual accuracy.
Extensive experiments demonstrate the proposed calibrated deep clustering model not only surpasses state-of-the-art deep clustering methods by 10 times in terms of expected calibration error but also significantly outperforms them in terms of clustering accuracy.
arXiv Detail & Related papers (2024-03-04T11:23:40Z) - Domain-adaptive and Subgroup-specific Cascaded Temperature Regression
for Out-of-distribution Calibration [16.930766717110053]
We propose a novel meta-set-based cascaded temperature regression method for post-hoc calibration.
We partition each meta-set into subgroups based on predicted category and confidence level, capturing diverse uncertainties.
A regression network is then trained to derive category-specific and confidence-level-specific scaling, achieving calibration across meta-sets.
arXiv Detail & Related papers (2024-02-14T14:35:57Z) - Calibration by Distribution Matching: Trainable Kernel Calibration
Metrics [56.629245030893685]
We introduce kernel-based calibration metrics that unify and generalize popular forms of calibration for both classification and regression.
These metrics admit differentiable sample estimates, making it easy to incorporate a calibration objective into empirical risk minimization.
We provide intuitive mechanisms to tailor calibration metrics to a decision task, and enforce accurate loss estimation and no regret decisions.
arXiv Detail & Related papers (2023-10-31T06:19:40Z) - Two Sides of Miscalibration: Identifying Over and Under-Confidence
Prediction for Network Calibration [1.192436948211501]
Proper confidence calibration of deep neural networks is essential for reliable predictions in safety-critical tasks.
Miscalibration can lead to model over-confidence and/or under-confidence.
We introduce a novel metric, a miscalibration score, to identify the overall and class-wise calibration status.
We use the class-wise miscalibration score as a proxy to design a calibration technique that can tackle both over and under-confidence.
arXiv Detail & Related papers (2023-08-06T17:59:14Z) - Calibration-Aware Bayesian Learning [37.82259435084825]
This paper proposes an integrated framework, referred to as calibration-aware Bayesian neural networks (CA-BNNs)
It applies both data-dependent or data-independent regularizers while optimizing over a variational distribution as in Bayesian learning.
Numerical results validate the advantages of the proposed approach in terms of expected calibration error (ECE) and reliability diagrams.
arXiv Detail & Related papers (2023-05-12T14:19:15Z) - Calibration of Neural Networks [77.34726150561087]
This paper presents a survey of confidence calibration problems in the context of neural networks.
We analyze problem statement, calibration definitions, and different approaches to evaluation.
Empirical experiments cover various datasets and models, comparing calibration methods according to different criteria.
arXiv Detail & Related papers (2023-03-19T20:27:51Z) - Better Uncertainty Calibration via Proper Scores for Classification and
Beyond [15.981380319863527]
We introduce the framework of proper calibration errors, which relates every calibration error to a proper score.
This relationship can be used to reliably quantify the model calibration improvement.
arXiv Detail & Related papers (2022-03-15T12:46:08Z) - Unsupervised Calibration under Covariate Shift [92.02278658443166]
We introduce the problem of calibration under domain shift and propose an importance sampling based approach to address it.
We evaluate and discuss the efficacy of our method on both real-world datasets and synthetic datasets.
arXiv Detail & Related papers (2020-06-29T21:50:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.