Uncertainty Quantification of the 4th kind; optimal posterior
accuracy-uncertainty tradeoff with the minimum enclosing ball
- URL: http://arxiv.org/abs/2108.10517v1
- Date: Tue, 24 Aug 2021 04:02:45 GMT
- Title: Uncertainty Quantification of the 4th kind; optimal posterior
accuracy-uncertainty tradeoff with the minimum enclosing ball
- Authors: Hamed Hamze Bajgiran and Pau Batlle Franch and Houman Owhadi and Clint
Scovel and Mahdy Shirdel and Michael Stanley and Peyman Tavallali
- Abstract summary: We introduce a 4th kind of approach to Uncertainty Quantification (UQ)
It can be summarized as, after observing a sample $x$, defining a likelihood region through the relative likelihood and playing a minmax game in that region to define optimal estimators and their risk.
The proposed method addresses the brittleness of Bayesian inference by navigating the robustness-accuracy tradeoff associated with data assimilation.
- Score: 1.6009195333398072
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There are essentially three kinds of approaches to Uncertainty Quantification
(UQ): (A) robust optimization, (B) Bayesian, (C) decision theory. Although (A)
is robust, it is unfavorable with respect to accuracy and data assimilation.
(B) requires a prior, it is generally brittle and posterior estimations can be
slow. Although (C) leads to the identification of an optimal prior, its
approximation suffers from the curse of dimensionality and the notion of risk
is one that is averaged with respect to the distribution of the data. We
introduce a 4th kind which is a hybrid between (A), (B), (C), and hypothesis
testing. It can be summarized as, after observing a sample $x$, (1) defining a
likelihood region through the relative likelihood and (2) playing a minmax game
in that region to define optimal estimators and their risk. The resulting
method has several desirable properties (a) an optimal prior is identified
after measuring the data, and the notion of risk is a posterior one, (b) the
determination of the optimal estimate and its risk can be reduced to computing
the minimum enclosing ball of the image of the likelihood region under the
quantity of interest map (which is fast and not subject to the curse of
dimensionality). The method is characterized by a parameter in $ [0,1]$ acting
as an assumed lower bound on the rarity of the observed data (the relative
likelihood). When that parameter is near $1$, the method produces a posterior
distribution concentrated around a maximum likelihood estimate with tight but
low confidence UQ estimates. When that parameter is near $0$, the method
produces a maximal risk posterior distribution with high confidence UQ
estimates. In addition to navigating the accuracy-uncertainty tradeoff, the
proposed method addresses the brittleness of Bayesian inference by navigating
the robustness-accuracy tradeoff associated with data assimilation.
Related papers
- Estimation of discrete distributions in relative entropy, and the deviations of the missing mass [3.4265828682659705]
We study the problem of estimating a distribution over a finite alphabet from an i.i.d. sample, with accuracy measured in relative entropy.
While optimal expected risk bounds are known, high-probability guarantees remain less well-understood.
arXiv Detail & Related papers (2025-04-30T16:47:10Z) - Exact characterization of ε-Safe Decision Regions for exponential family distributions and Multi Cost SVM approximation [2.225810431340323]
We introduce a formal definition of epsilon-Safe Decision Region, a subset of the input space in which the prediction of a target (safe) class is probabilistically guaranteed.
Inspired by this limitation, we developed Multi Cost SVM, an SVM based algorithm that approximates the safe region and is also able to handle unbalanced data.
arXiv Detail & Related papers (2025-01-29T16:14:35Z) - Data-driven decision-making under uncertainty with entropic risk measure [5.407319151576265]
The entropic risk measure is widely used in high-stakes decision making to account for tail risks associated with an uncertain loss.
To debias the empirical entropic risk estimator, we propose a strongly consistent bootstrapping procedure.
We show that cross validation methods can result in significantly higher out-of-sample risk for the insurer if the bias in validation performance is not corrected for.
arXiv Detail & Related papers (2024-09-30T04:02:52Z) - Information-Theoretic Safe Bayesian Optimization [59.758009422067005]
We consider a sequential decision making task, where the goal is to optimize an unknown function without evaluating parameters that violate an unknown (safety) constraint.
Most current methods rely on a discretization of the domain and cannot be directly extended to the continuous case.
We propose an information-theoretic safe exploration criterion that directly exploits the GP posterior to identify the most informative safe parameters to evaluate.
arXiv Detail & Related papers (2024-02-23T14:31:10Z) - Calibrating Neural Simulation-Based Inference with Differentiable
Coverage Probability [50.44439018155837]
We propose to include a calibration term directly into the training objective of the neural model.
By introducing a relaxation of the classical formulation of calibration error we enable end-to-end backpropagation.
It is directly applicable to existing computational pipelines allowing reliable black-box posterior inference.
arXiv Detail & Related papers (2023-10-20T10:20:45Z) - A Mean Field Approach to Empirical Bayes Estimation in High-dimensional
Linear Regression [8.345523969593492]
We study empirical Bayes estimation in high-dimensional linear regression.
We adopt a variational empirical Bayes approach, introduced originally in Carbonetto and Stephens (2012) and Kim et al. (2022).
This provides the first rigorous empirical Bayes method in a high-dimensional regression setting without sparsity.
arXiv Detail & Related papers (2023-09-28T20:51:40Z) - Minimax Instrumental Variable Regression and $L_2$ Convergence
Guarantees without Identification or Closedness [71.42652863687117]
We study nonparametric estimation of instrumental variable (IV) regressions.
We propose a new penalized minimax estimator that can converge to a fixed IV solution.
We derive a strong $L$ error rate for our estimator under lax conditions.
arXiv Detail & Related papers (2023-02-10T18:08:49Z) - Bounding Box-based Multi-objective Bayesian Optimization of Risk
Measures under Input Uncertainty [21.056363101777052]
We propose a novel multi-objective Bayesian optimization (MOBO) method to efficiently identify the Pareto front (PF) defined by risk measures for black-box functions under the presence of input uncertainty (IU)
The basic idea of the proposed method is to assume a Gaussian process (GP) model for the black-box function and to construct high-probability bounding boxes for the risk measures using the GP model.
We prove that the algorithm can return an arbitrary-accurate solution in a finite number of iterations with high probability, for various risk measures such as Bayes risk, worst-case risk, and value-
arXiv Detail & Related papers (2023-01-27T08:25:45Z) - A Dimensionality Reduction Method for Finding Least Favorable Priors
with a Focus on Bregman Divergence [108.28566246421742]
This paper develops a dimensionality reduction method that allows us to move the optimization to a finite-dimensional setting with an explicit bound on the dimension.
In order to make progress on the problem, we restrict ourselves to Bayesian risks induced by a relatively large class of loss functions, namely Bregman divergences.
arXiv Detail & Related papers (2022-02-23T16:22:28Z) - Tractable and Near-Optimal Adversarial Algorithms for Robust Estimation
in Contaminated Gaussian Models [1.609950046042424]
Consider the problem of simultaneous estimation of location and variance matrix under Huber's contaminated Gaussian model.
First, we study minimum $f$-divergence estimation at the population level, corresponding to a generative adversarial method with a nonparametric discriminator.
We develop tractable adversarial algorithms with simple spline discriminators, which can be implemented via nested optimization.
The proposed methods are shown to achieve minimax optimal rates or near-optimal rates depending on the $f$-divergence and the penalty used.
arXiv Detail & Related papers (2021-12-24T02:46:51Z) - Amortized Conditional Normalized Maximum Likelihood: Reliable Out of
Distribution Uncertainty Estimation [99.92568326314667]
We propose the amortized conditional normalized maximum likelihood (ACNML) method as a scalable general-purpose approach for uncertainty estimation.
Our algorithm builds on the conditional normalized maximum likelihood (CNML) coding scheme, which has minimax optimal properties according to the minimum description length principle.
We demonstrate that ACNML compares favorably to a number of prior techniques for uncertainty estimation in terms of calibration on out-of-distribution inputs.
arXiv Detail & Related papers (2020-11-05T08:04:34Z) - $\gamma$-ABC: Outlier-Robust Approximate Bayesian Computation Based on a
Robust Divergence Estimator [95.71091446753414]
We propose to use a nearest-neighbor-based $gamma$-divergence estimator as a data discrepancy measure.
Our method achieves significantly higher robustness than existing discrepancy measures.
arXiv Detail & Related papers (2020-06-13T06:09:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.