Distribution-Free Inference for the Regression Function of Binary
Classification
- URL: http://arxiv.org/abs/2308.01835v1
- Date: Thu, 3 Aug 2023 15:52:27 GMT
- Title: Distribution-Free Inference for the Regression Function of Binary
Classification
- Authors: Ambrus Tam\'as and Bal\'azs Csan\'ad Cs\'aji
- Abstract summary: The paper presents a resampling framework to construct exact, distribution-free and non-asymptotically guaranteed confidence regions for the true regression function for any user-chosen confidence level.
It is proved that the constructed confidence regions are strongly consistent, that is, any false model is excluded in the long run with probability one.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One of the key objects of binary classification is the regression function,
i.e., the conditional expectation of the class labels given the inputs. With
the regression function not only a Bayes optimal classifier can be defined, but
it also encodes the corresponding misclassification probabilities. The paper
presents a resampling framework to construct exact, distribution-free and
non-asymptotically guaranteed confidence regions for the true regression
function for any user-chosen confidence level. Then, specific algorithms are
suggested to demonstrate the framework. It is proved that the constructed
confidence regions are strongly consistent, that is, any false model is
excluded in the long run with probability one. The exclusion is quantified with
probably approximately correct type bounds, as well. Finally, the algorithms
are validated via numerical experiments, and the methods are compared to
approximate asymptotic confidence ellipsoids.
Related papers
- Calibrating Neural Simulation-Based Inference with Differentiable
Coverage Probability [50.44439018155837]
We propose to include a calibration term directly into the training objective of the neural model.
By introducing a relaxation of the classical formulation of calibration error we enable end-to-end backpropagation.
It is directly applicable to existing computational pipelines allowing reliable black-box posterior inference.
arXiv Detail & Related papers (2023-10-20T10:20:45Z) - When Does Confidence-Based Cascade Deferral Suffice? [69.28314307469381]
Cascades are a classical strategy to enable inference cost to vary adaptively across samples.
A deferral rule determines whether to invoke the next classifier in the sequence, or to terminate prediction.
Despite being oblivious to the structure of the cascade, confidence-based deferral often works remarkably well in practice.
arXiv Detail & Related papers (2023-07-06T04:13:57Z) - Multivariate Probabilistic Regression with Natural Gradient Boosting [63.58097881421937]
We propose a Natural Gradient Boosting (NGBoost) approach based on nonparametrically modeling the conditional parameters of the multivariate predictive distribution.
Our method is robust, works out-of-the-box without extensive tuning, is modular with respect to the assumed target distribution, and performs competitively in comparison to existing approaches.
arXiv Detail & Related papers (2021-06-07T17:44:49Z) - Root-finding Approaches for Computing Conformal Prediction Set [18.405645120971496]
Conformal prediction constructs a confidence region for an unobserved response of a feature vector based on previous identically distributed and exchangeable observations.
We exploit the fact that, emphoften, conformal prediction sets are intervals whose boundaries can be efficiently approximated by classical root-finding software.
arXiv Detail & Related papers (2021-04-14T06:41:12Z) - Learning Probabilistic Ordinal Embeddings for Uncertainty-Aware
Regression [91.3373131262391]
Uncertainty is the only certainty there is.
Traditionally, the direct regression formulation is considered and the uncertainty is modeled by modifying the output space to a certain family of probabilistic distributions.
How to model the uncertainty within the present-day technologies for regression remains an open issue.
arXiv Detail & Related papers (2021-03-25T06:56:09Z) - Exact Distribution-Free Hypothesis Tests for the Regression Function of
Binary Classification via Conditional Kernel Mean Embeddings [0.0]
Two hypothesis tests are proposed for the regression function of binary classification based on conditional kernel mean embeddings.
Tests are introduced in a flexible manner allowing us to control the exact probability of type I error.
arXiv Detail & Related papers (2021-03-08T22:31:23Z) - Estimation and Applications of Quantiles in Deep Binary Classification [0.0]
Quantile regression, based on check loss, is a widely used inferential paradigm in Statistics.
We consider the analogue of check loss in the binary classification setting.
We develop individualized confidence scores that can be used to decide whether a prediction is reliable.
arXiv Detail & Related papers (2021-02-09T07:07:42Z) - Optimal strategies for reject option classifiers [0.0]
In classification with a reject option, the classifier is allowed in uncertain cases to abstain from prediction.
We coin a symmetric definition, the bounded-coverage model, which seeks for a classifier with minimal selective risk and guaranteed coverage.
We propose two algorithms to learn the proper uncertainty score from examples for an arbitrary black-box classifier.
arXiv Detail & Related papers (2021-01-29T11:09:32Z) - Certifying Confidence via Randomized Smoothing [151.67113334248464]
Randomized smoothing has been shown to provide good certified-robustness guarantees for high-dimensional classification problems.
Most smoothing methods do not give us any information about the confidence with which the underlying classifier makes a prediction.
We propose a method to generate certified radii for the prediction confidence of the smoothed classifier.
arXiv Detail & Related papers (2020-09-17T04:37:26Z) - Evaluating probabilistic classifiers: Reliability diagrams and score
decompositions revisited [68.8204255655161]
We introduce the CORP approach, which generates provably statistically Consistent, Optimally binned, and Reproducible reliability diagrams in an automated way.
Corpor is based on non-parametric isotonic regression and implemented via the Pool-adjacent-violators (PAV) algorithm.
arXiv Detail & Related papers (2020-08-07T08:22:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.