A novel Deep Learning approach for one-step Conformal Prediction
approximation
- URL: http://arxiv.org/abs/2207.12377v4
- Date: Mon, 7 Aug 2023 12:02:40 GMT
- Title: A novel Deep Learning approach for one-step Conformal Prediction
approximation
- Authors: Julia A. Meister, Khuong An Nguyen, Stelios Kapetanakis, Zhiyuan Luo
- Abstract summary: Conformal Prediction (CP) is a versatile solution that guarantees a maximum error rate given minimal constraints.
We propose a novel conformal loss function that approximates the traditionally two-step CP approach in a single step.
- Score: 0.7646713951724009
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Deep Learning predictions with measurable confidence are increasingly
desirable for real-world problems, especially in high-risk settings. The
Conformal Prediction (CP) framework is a versatile solution that guarantees a
maximum error rate given minimal constraints. In this paper, we propose a novel
conformal loss function that approximates the traditionally two-step CP
approach in a single step. By evaluating and penalising deviations from the
stringent expected CP output distribution, a Deep Learning model may learn the
direct relationship between the input data and the conformal p-values. We carry
out a comprehensive empirical evaluation to show our novel loss function's
competitiveness for seven binary and multi-class prediction tasks on five
benchmark datasets. On the same datasets, our approach achieves significant
training time reductions up to 86% compared to Aggregated Conformal Prediction
(ACP), while maintaining comparable approximate validity and predictive
efficiency.
Related papers
- Conformal Risk Minimization with Variance Reduction [37.74931189657469]
Conformal prediction (CP) is a distribution-free framework for achieving probabilistic guarantees on black-box models.
Recent research efforts have focused on optimizing CP efficiency during training.
We formalize this concept as the problem of conformal risk minimization.
arXiv Detail & Related papers (2024-11-03T21:48:15Z) - Conformal Thresholded Intervals for Efficient Regression [9.559062601251464]
Conformal Thresholded Intervals (CTI) is a novel conformal regression method that aims to produce the smallest possible prediction set with guaranteed coverage.
CTI constructs prediction sets by thresholding the estimated conditional interquantile intervals based on their length.
CTI achieves superior performance compared to state-of-the-art conformal regression methods across various datasets.
arXiv Detail & Related papers (2024-07-19T17:47:08Z) - Adapting Conformal Prediction to Distribution Shifts Without Labels [16.478151550456804]
Conformal prediction (CP) enables machine learning models to output prediction sets with guaranteed coverage rate.
Our goal is to improve the quality of CP-generated prediction sets using only unlabeled data from the test domain.
This is achieved by two new methods called ECP and EACP, that adjust the score function in CP according to the base model's uncertainty on the unlabeled test data.
arXiv Detail & Related papers (2024-06-03T15:16:02Z) - Loss Shaping Constraints for Long-Term Time Series Forecasting [79.3533114027664]
We present a Constrained Learning approach for long-term time series forecasting that respects a user-defined upper bound on the loss at each time-step.
We propose a practical Primal-Dual algorithm to tackle it, and aims to demonstrate that it exhibits competitive average performance in time series benchmarks, while shaping the errors across the predicted window.
arXiv Detail & Related papers (2024-02-14T18:20:44Z) - Efficient and Differentiable Conformal Prediction with General Function
Classes [96.74055810115456]
We propose a generalization of conformal prediction to multiple learnable parameters.
We show that it achieves approximate valid population coverage and near-optimal efficiency within class.
Experiments show that our algorithm is able to learn valid prediction sets and improve the efficiency significantly.
arXiv Detail & Related papers (2022-02-22T18:37:23Z) - Adaptive Affinity Loss and Erroneous Pseudo-Label Refinement for Weakly
Supervised Semantic Segmentation [48.294903659573585]
In this paper, we propose to embed affinity learning of multi-stage approaches in a single-stage model.
A deep neural network is used to deliver comprehensive semantic information in the training phase.
Experiments are conducted on the PASCAL VOC 2012 dataset to evaluate the effectiveness of our proposed approach.
arXiv Detail & Related papers (2021-08-03T07:48:33Z) - Learning Prediction Intervals for Regression: Generalization and
Calibration [12.576284277353606]
We study the generation of prediction intervals in regression for uncertainty quantification.
We use a general learning theory to characterize the optimality-feasibility tradeoff that encompasses Lipschitz continuity and VC-subgraph classes.
We empirically demonstrate the strengths of our interval generation and calibration algorithms in terms of testing performances compared to existing benchmarks.
arXiv Detail & Related papers (2021-02-26T17:55:30Z) - Doubly Robust Off-Policy Actor-Critic: Convergence and Optimality [131.45028999325797]
We develop a doubly robust off-policy AC (DR-Off-PAC) for discounted MDP.
DR-Off-PAC adopts a single timescale structure, in which both actor and critics are updated simultaneously with constant stepsize.
We study the finite-time convergence rate and characterize the sample complexity for DR-Off-PAC to attain an $epsilon$-accurate optimal policy.
arXiv Detail & Related papers (2021-02-23T18:56:13Z) - CoinDICE: Off-Policy Confidence Interval Estimation [107.86876722777535]
We study high-confidence behavior-agnostic off-policy evaluation in reinforcement learning.
We show in a variety of benchmarks that the confidence interval estimates are tighter and more accurate than existing methods.
arXiv Detail & Related papers (2020-10-22T12:39:11Z) - Evaluating probabilistic classifiers: Reliability diagrams and score
decompositions revisited [68.8204255655161]
We introduce the CORP approach, which generates provably statistically Consistent, Optimally binned, and Reproducible reliability diagrams in an automated way.
Corpor is based on non-parametric isotonic regression and implemented via the Pool-adjacent-violators (PAV) algorithm.
arXiv Detail & Related papers (2020-08-07T08:22:26Z) - PIVEN: A Deep Neural Network for Prediction Intervals with Specific
Value Prediction [14.635820704895034]
We present PIVEN, a deep neural network for producing both a PI and a value prediction.
Our approach makes no assumptions regarding data distribution within the PI, making its value prediction more effective for various real-world problems.
arXiv Detail & Related papers (2020-06-09T09:29:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.