T-SCI: A Two-Stage Conformal Inference Algorithm with Guaranteed
Coverage for Cox-MLP
- URL: http://arxiv.org/abs/2103.04556v1
- Date: Mon, 8 Mar 2021 05:42:05 GMT
- Title: T-SCI: A Two-Stage Conformal Inference Algorithm with Guaranteed
Coverage for Cox-MLP
- Authors: Jiaye Teng, Zeren Tan, Yang Yuan
- Abstract summary: We propose two algorithms for recovering guaranteed coverage in censored data.
First, we revisit weighted conformal inference and introduce a new non-conformity score based on partial likelihood.
We then propose a two-stage algorithm emphT-SCI, where we run WCCI in the first stage and apply quantile conformal inference to calibrate the results in the second stage.
- Score: 13.4379473119565
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It is challenging to deal with censored data, where we only have access to
the incomplete information of survival time instead of its exact value.
Fortunately, under linear predictor assumption, people can obtain guaranteed
coverage for the confidence band of survival time using methods like Cox
Regression. However, when relaxing the linear assumption with neural networks
(e.g., Cox-MLP \citep{katzman2018deepsurv,kvamme2019time}), we lose the
guaranteed coverage. To recover the guaranteed coverage without linear
assumption, we propose two algorithms based on conformal inference. In the
first algorithm \emph{WCCI}, we revisit weighted conformal inference and
introduce a new non-conformity score based on partial likelihood. We then
propose a two-stage algorithm \emph{T-SCI}, where we run WCCI in the first
stage and apply quantile conformal inference to calibrate the results in the
second stage. Theoretical analysis shows that T-SCI returns guaranteed coverage
under milder assumptions than WCCI. We conduct extensive experiments on
synthetic data and real data using different methods, which validate our
analysis.
Related papers
- The Benefit of Being Bayesian in Online Conformal Prediction [7.713245413733777]
We study the online construction of valid confidence sets given a black-box machine learning model.
By converting the target confidence levels into quantile levels, the problem can be reduced to predicting the quantiles of a sequentially revealed data sequence.
arXiv Detail & Related papers (2024-10-03T15:04:47Z) - Provably Robust Conformal Prediction with Improved Efficiency [29.70455766394585]
Conformal prediction is a powerful tool to generate uncertainty sets with guaranteed coverage.
adversarial examples are able to manipulate conformal methods to construct prediction sets with invalid coverage rates.
We propose two novel methods, Post-Training Transformation (PTT) and Robust Conformal Training (RCT), to effectively reduce prediction set size with little overhead.
arXiv Detail & Related papers (2024-04-30T15:49:01Z) - Equal Opportunity of Coverage in Fair Regression [50.76908018786335]
We study fair machine learning (ML) under predictive uncertainty to enable reliable and trustworthy decision-making.
We propose Equal Opportunity of Coverage (EOC) that aims to achieve two properties: (1) coverage rates for different groups with similar outcomes are close, and (2) the coverage rate for the entire population remains at a predetermined level.
arXiv Detail & Related papers (2023-11-03T21:19:59Z) - Calibrating Neural Simulation-Based Inference with Differentiable
Coverage Probability [50.44439018155837]
We propose to include a calibration term directly into the training objective of the neural model.
By introducing a relaxation of the classical formulation of calibration error we enable end-to-end backpropagation.
It is directly applicable to existing computational pipelines allowing reliable black-box posterior inference.
arXiv Detail & Related papers (2023-10-20T10:20:45Z) - Online Heavy-tailed Change-point detection [6.7643284029102295]
We present an algorithm based on clipped Gradient Descent (SGD), that works even if we only assume that the second moment of the data generating process is bounded.
We derive guarantees on worst-case, finite-sample false-positive rate (FPR) over the family of all distributions with bounded second moment.
Our method is the first OCPD algorithm that guarantees finite-sample FPR, even if the data is high dimensional and the underlying distributions are heavy-tailed.
arXiv Detail & Related papers (2023-06-15T23:39:05Z) - Fully Stochastic Trust-Region Sequential Quadratic Programming for
Equality-Constrained Optimization Problems [62.83783246648714]
We propose a sequential quadratic programming algorithm (TR-StoSQP) to solve nonlinear optimization problems with objectives and deterministic equality constraints.
The algorithm adaptively selects the trust-region radius and, compared to the existing line-search StoSQP schemes, allows us to utilize indefinite Hessian matrices.
arXiv Detail & Related papers (2022-11-29T05:52:17Z) - Risk Minimization from Adaptively Collected Data: Guarantees for
Supervised and Policy Learning [57.88785630755165]
Empirical risk minimization (ERM) is the workhorse of machine learning, but its model-agnostic guarantees can fail when we use adaptively collected data.
We study a generic importance sampling weighted ERM algorithm for using adaptively collected data to minimize the average of a loss function over a hypothesis class.
For policy learning, we provide rate-optimal regret guarantees that close an open gap in the existing literature whenever exploration decays to zero.
arXiv Detail & Related papers (2021-06-03T09:50:13Z) - Large-Scale Methods for Distributionally Robust Optimization [53.98643772533416]
We prove that our algorithms require a number of evaluations gradient independent of training set size and number of parameters.
Experiments on MNIST and ImageNet confirm the theoretical scaling of our algorithms, which are 9--36 times more efficient than full-batch methods.
arXiv Detail & Related papers (2020-10-12T17:41:44Z) - Fast OSCAR and OWL Regression via Safe Screening Rules [97.28167655721766]
Ordered $L_1$ (OWL) regularized regression is a new regression analysis for high-dimensional sparse learning.
Proximal gradient methods are used as standard approaches to solve OWL regression.
We propose the first safe screening rule for OWL regression by exploring the order of the primal solution with the unknown order structure.
arXiv Detail & Related papers (2020-06-29T23:35:53Z) - Provable Training of a ReLU Gate with an Iterative Non-Gradient
Algorithm [0.7614628596146599]
We show provable guarantees on the training of a single ReLU gate in hitherto unexplored regimes.
We show a first-of-its-kind approximate recovery of the true label generating parameters under an (online) data-poisoning attack on the true labels.
Our guarantee is shown to be nearly optimal in the worst case and its accuracy of recovering the true weight degrades gracefully with increasing probability of attack and its magnitude.
arXiv Detail & Related papers (2020-05-08T17:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.