Conformalized Online Learning: Online Calibration Without a Holdout Set
- URL: http://arxiv.org/abs/2205.09095v2
- Date: Thu, 19 May 2022 17:56:57 GMT
- Title: Conformalized Online Learning: Online Calibration Without a Holdout Set
- Authors: Shai Feldman, Stephen Bates, Yaniv Romano
- Abstract summary: We develop a framework for constructing uncertainty sets with a valid coverage guarantee in an online setting.
We show how to construct valid intervals for a multiple-output regression problem.
- Score: 10.420394952839242
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We develop a framework for constructing uncertainty sets with a valid
coverage guarantee in an online setting, in which the underlying data
distribution can drastically -- and even adversarially -- shift over time. The
technique we propose is highly flexible as it can be integrated with any online
learning algorithm, requiring minimal implementation effort and computational
cost. A key advantage of our method over existing alternatives -- which also
build on conformal inference -- is that we do not need to split the data into
training and holdout calibration sets. This allows us to fit the predictive
model in a fully online manner, utilizing the most recent observation for
constructing calibrated uncertainty sets. Consequently, and in contrast with
existing techniques, (i) the sets we build can quickly adapt to new changes in
the distribution; and (ii) our procedure does not require refitting the model
at each time step. Using synthetic and real-world benchmark data sets, we
demonstrate the validity of our theory and the improved performance of our
proposal over existing techniques. To demonstrate the greater flexibility of
the proposed method, we show how to construct valid intervals for a
multiple-output regression problem that previous sequential calibration methods
cannot handle due to impractical computational and memory requirements.
Related papers
- Towards Stable Machine Learning Model Retraining via Slowly Varying Sequences [6.067007470552307]
We propose a methodology for finding sequences of machine learning models that are stable across retraining iterations.
We develop a mixed-integer optimization formulation that is guaranteed to recover optimal models.
Our method shows stronger stability than greedily trained models with a small, controllable sacrifice in predictive power.
arXiv Detail & Related papers (2024-03-28T22:45:38Z) - Improving Adaptive Conformal Prediction Using Self-Supervised Learning [72.2614468437919]
We train an auxiliary model with a self-supervised pretext task on top of an existing predictive model and use the self-supervised error as an additional feature to estimate nonconformity scores.
We empirically demonstrate the benefit of the additional information using both synthetic and real data on the efficiency (width), deficit, and excess of conformal prediction intervals.
arXiv Detail & Related papers (2023-02-23T18:57:14Z) - Improved Online Conformal Prediction via Strongly Adaptive Online
Learning [86.4346936885507]
We develop new online conformal prediction methods that minimize the strongly adaptive regret.
We prove that our methods achieve near-optimal strongly adaptive regret for all interval lengths simultaneously.
Experiments show that our methods consistently obtain better coverage and smaller prediction sets than existing methods on real-world tasks.
arXiv Detail & Related papers (2023-02-15T18:59:30Z) - Conformal Inference for Online Prediction with Arbitrary Distribution
Shifts [1.2277343096128712]
We consider the problem of forming prediction sets in an online setting where the distribution generating the data is allowed to vary over time.
We develop a novel procedure with provably small regret over all local time intervals of a given width.
We test our techniques on two real-world datasets aimed at predicting stock market volatility and COVID-19 case counts.
arXiv Detail & Related papers (2022-08-17T16:51:12Z) - Modular Conformal Calibration [80.33410096908872]
We introduce a versatile class of algorithms for recalibration in regression.
This framework allows one to transform any regression model into a calibrated probabilistic model.
We conduct an empirical study of MCC on 17 regression datasets.
arXiv Detail & Related papers (2022-06-23T03:25:23Z) - HyperImpute: Generalized Iterative Imputation with Automatic Model
Selection [77.86861638371926]
We propose a generalized iterative imputation framework for adaptively and automatically configuring column-wise models.
We provide a concrete implementation with out-of-the-box learners, simulators, and interfaces.
arXiv Detail & Related papers (2022-06-15T19:10:35Z) - Model-based Offline Imitation Learning with Non-expert Data [7.615595533111191]
We propose a scalable model-based offline imitation learning algorithmic framework that leverages datasets collected by both suboptimal and optimal policies.
We show that the proposed method textitalways outperforms Behavioral Cloning in the low data regime on simulated continuous control domains.
arXiv Detail & Related papers (2022-06-11T13:08:08Z) - A Regularized Implicit Policy for Offline Reinforcement Learning [54.7427227775581]
offline reinforcement learning enables learning from a fixed dataset, without further interactions with the environment.
We propose a framework that supports learning a flexible yet well-regularized fully-implicit policy.
Experiments and ablation study on the D4RL dataset validate our framework and the effectiveness of our algorithmic designs.
arXiv Detail & Related papers (2022-02-19T20:22:04Z) - An Adaptive Framework for Learning Unsupervised Depth Completion [59.17364202590475]
We present a method to infer a dense depth map from a color image and associated sparse depth measurements.
We show that regularization and co-visibility are related via the fitness of the model to data and can be unified into a single framework.
arXiv Detail & Related papers (2021-06-06T02:27:55Z) - Online Learning Based Risk-Averse Stochastic MPC of Constrained Linear
Uncertain Systems [0.0]
This paper investigates the problem of designing data-driven Model Predictive Control (MPC) for linear time-invariant systems under additive disturbance.
We propose a novel online learning based risk-varying MPC framework in which Conditional Value-at-Risk (CVaR) constraints are required to hold for a family of distributions called an ambiguity set.
arXiv Detail & Related papers (2020-11-20T13:00:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.