Bayes in Wonderland! Predictive supervised classification inference hits
unpredictability
- URL: http://arxiv.org/abs/2112.01880v1
- Date: Fri, 3 Dec 2021 12:34:52 GMT
- Title: Bayes in Wonderland! Predictive supervised classification inference hits
unpredictability
- Authors: Ali Amiryousefi, Ville Kinnula, Jing Tang
- Abstract summary: We show the convergence of the sBpc and mBpc under de Finetti type of exchangeability.
We also provide a parameter estimation of the generative model giving rise to the partition exchangeable sequence.
- Score: 1.8814209805277506
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The marginal Bayesian predictive classifiers (mBpc) as opposed to the
simultaneous Bayesian predictive classifiers (sBpc), handle each data
separately and hence tacitly assumes the independence of the observations.
However, due to saturation in learning of generative model parameters, the
adverse effect of this false assumption on the accuracy of mBpc tends to wear
out in face of increasing amount of training data; guaranteeing the convergence
of these two classifiers under de Finetti type of exchangeability. This result
however, is far from trivial for the sequences generated under Partition
exchangeability (PE), where even umpteen amount of training data is not ruling
out the possibility of an unobserved outcome (Wonderland!). We provide a
computational scheme that allows the generation of the sequences under PE.
Based on that, with controlled increase of the training data, we show the
convergence of the sBpc and mBpc. This underlies the use of simpler yet
computationally more efficient marginal classifiers instead of simultaneous. We
also provide a parameter estimation of the generative model giving rise to the
partition exchangeable sequence as well as a testing paradigm for the equality
of this parameter across different samples. The package for Bayesian predictive
supervised classifications, parameter estimation and hypothesis testing of the
Ewens Sampling formula generative model is deposited on CRAN as PEkit package
and free available from https://github.com/AmiryousefiLab/PEkit.
Related papers
- Split Conformal Prediction under Data Contamination [14.23965125128232]
We study the robustness of split conformal prediction in a data contamination setting.
We quantify the impact of corrupted data on the coverage and efficiency of the constructed sets.
We propose an adjustment in the classification setting which we call Contamination Robust Conformal Prediction.
arXiv Detail & Related papers (2024-07-10T14:33:28Z) - Posterior Uncertainty Quantification in Neural Networks using Data Augmentation [3.9860047080844807]
We show that deep ensembling is a fundamentally mis-specified model class, since it assumes that future data are supported on existing observations only.
We propose MixupMP, a method that constructs a more realistic predictive distribution using popular data augmentation techniques.
Our empirical analysis showcases that MixupMP achieves superior predictive performance and uncertainty quantification on various image classification datasets.
arXiv Detail & Related papers (2024-03-18T17:46:07Z) - Probabilistic Contrastive Learning for Long-Tailed Visual Recognition [78.70453964041718]
Longtailed distributions frequently emerge in real-world data, where a large number of minority categories contain a limited number of samples.
Recent investigations have revealed that supervised contrastive learning exhibits promising potential in alleviating the data imbalance.
We propose a novel probabilistic contrastive (ProCo) learning algorithm that estimates the data distribution of the samples from each class in the feature space.
arXiv Detail & Related papers (2024-03-11T13:44:49Z) - Bayesian Quantification with Black-Box Estimators [1.599072005190786]
Approaches like adjusted classify and count, black-box shift estimators, and invariant ratio estimators use an auxiliary (and potentially biased) black-box classifier to estimate the class distribution and yield guarantees under weak assumptions.
We demonstrate that all these algorithms are closely related to the inference in a particular Bayesian Chain model, approxing the assumed ground-truthgenerative process.
Then, we discuss an efficient Markov Monte Carlo sampling scheme for the introduced model and show an consistency guarantee in the large-data limit.
arXiv Detail & Related papers (2023-02-17T22:10:04Z) - A Bayesian Semiparametric Method For Estimating Causal Quantile Effects [1.1118668841431563]
We propose a semiparametric conditional distribution regression model that allows inference on any functionals of counterfactual distributions.
We show via simulations that the use of double balancing score for confounding adjustment improves performance over adjusting for any single score alone.
We apply the proposed method to the North Carolina birth weight dataset to analyze the effect of maternal smoking on infant's birth weight.
arXiv Detail & Related papers (2022-11-03T05:15:18Z) - Probabilistic Conformal Prediction Using Conditional Random Samples [73.26753677005331]
PCP is a predictive inference algorithm that estimates a target variable by a discontinuous predictive set.
It is efficient and compatible with either explicit or implicit conditional generative models.
arXiv Detail & Related papers (2022-06-14T03:58:03Z) - Self-Certifying Classification by Linearized Deep Assignment [65.0100925582087]
We propose a novel class of deep predictors for classifying metric data on graphs within PAC-Bayes risk certification paradigm.
Building on the recent PAC-Bayes literature and data-dependent priors, this approach enables learning posterior distributions on the hypothesis space.
arXiv Detail & Related papers (2022-01-26T19:59:14Z) - Transformers Can Do Bayesian Inference [56.99390658880008]
We present Prior-Data Fitted Networks (PFNs)
PFNs leverage in-context learning in large-scale machine learning techniques to approximate a large set of posteriors.
We demonstrate that PFNs can near-perfectly mimic Gaussian processes and also enable efficient Bayesian inference for intractable problems.
arXiv Detail & Related papers (2021-12-20T13:07:39Z) - Training on Test Data with Bayesian Adaptation for Covariate Shift [96.3250517412545]
Deep neural networks often make inaccurate predictions with unreliable uncertainty estimates.
We derive a Bayesian model that provides for a well-defined relationship between unlabeled inputs under distributional shift and model parameters.
We show that our method improves both accuracy and uncertainty estimation.
arXiv Detail & Related papers (2021-09-27T01:09:08Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.