Sharp Deviations Bounds for Dirichlet Weighted Sums with Application to
analysis of Bayesian algorithms
- URL: http://arxiv.org/abs/2304.03056v1
- Date: Thu, 6 Apr 2023 13:15:51 GMT
- Title: Sharp Deviations Bounds for Dirichlet Weighted Sums with Application to
analysis of Bayesian algorithms
- Authors: Denis Belomestny, Pierre Menard, Alexey Naumov, Daniil Tiapkin, Michal
Valko
- Abstract summary: We derive sharp non-asymptotic deviation bounds for weighted sums of Dirichlet random variables.
These bounds are based on a novel integral representation of the density of a weighted Dirichlet sum.
- Score: 30.310960863405267
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this work, we derive sharp non-asymptotic deviation bounds for weighted
sums of Dirichlet random variables. These bounds are based on a novel integral
representation of the density of a weighted Dirichlet sum. This representation
allows us to obtain a Gaussian-like approximation for the sum distribution
using geometry and complex analysis methods. Our results generalize similar
bounds for the Beta distribution obtained in the seminal paper Alfers and
Dinges [1984]. Additionally, our results can be considered a sharp
non-asymptotic version of the inverse of Sanov's theorem studied by Ganesh and
O'Connell [1999] in the Bayesian setting. Based on these results, we derive new
deviation bounds for the Dirichlet process posterior means with application to
Bayesian bootstrap. Finally, we apply our estimates to the analysis of the
Multinomial Thompson Sampling (TS) algorithm in multi-armed bandits and
significantly sharpen the existing regret bounds by making them independent of
the size of the arms distribution support.
Related papers
- Distributed Markov Chain Monte Carlo Sampling based on the Alternating
Direction Method of Multipliers [143.6249073384419]
In this paper, we propose a distributed sampling scheme based on the alternating direction method of multipliers.
We provide both theoretical guarantees of our algorithm's convergence and experimental evidence of its superiority to the state-of-the-art.
In simulation, we deploy our algorithm on linear and logistic regression tasks and illustrate its fast convergence compared to existing gradient-based methods.
arXiv Detail & Related papers (2024-01-29T02:08:40Z) - Intrinsic Bayesian Cramér-Rao Bound with an Application to Covariance Matrix Estimation [49.67011673289242]
This paper presents a new performance bound for estimation problems where the parameter to estimate lies in a smooth manifold.
It induces a geometry for the parameter manifold, as well as an intrinsic notion of the estimation error measure.
arXiv Detail & Related papers (2023-11-08T15:17:13Z) - A General Recipe for the Analysis of Randomized Multi-Armed Bandit
Algorithms [16.114012813668932]
We revisit two famous bandit algorithms, Minimum Empirical Divergence (MED) and Thompson Sampling (TS)
We prove that MED is optimal for all these models, but also provide a simple regret analysis of some TS algorithms for which the optimality is already known.
arXiv Detail & Related papers (2023-03-10T16:43:48Z) - Mean-Square Analysis of Discretized It\^o Diffusions for Heavy-tailed
Sampling [17.415391025051434]
We analyze the complexity of sampling from a class of heavy-tailed distributions by discretizing a natural class of Ito diffusions associated with weighted Poincar'e inequalities.
Based on a mean-square analysis, we establish the iteration complexity for obtaining a sample whose distribution is $epsilon$ close to the target distribution in the Wasserstein-2 metric.
arXiv Detail & Related papers (2023-03-01T15:16:03Z) - Robust Estimation for Nonparametric Families via Generative Adversarial
Networks [92.64483100338724]
We provide a framework for designing Generative Adversarial Networks (GANs) to solve high dimensional robust statistics problems.
Our work extend these to robust mean estimation, second moment estimation, and robust linear regression.
In terms of techniques, our proposed GAN losses can be viewed as a smoothed and generalized Kolmogorov-Smirnov distance.
arXiv Detail & Related papers (2022-02-02T20:11:33Z) - Estimating Gaussian Copulas with Missing Data [0.0]
We show how to circumvent a priori assumptions on the marginals with semiparametric modelling.
The joint distribution learned through this algorithm is considerably closer to the underlying distribution than existing methods.
arXiv Detail & Related papers (2022-01-14T17:20:44Z) - Pathwise Conditioning of Gaussian Processes [72.61885354624604]
Conventional approaches for simulating Gaussian process posteriors view samples as draws from marginal distributions of process values at finite sets of input locations.
This distribution-centric characterization leads to generative strategies that scale cubically in the size of the desired random vector.
We show how this pathwise interpretation of conditioning gives rise to a general family of approximations that lend themselves to efficiently sampling Gaussian process posteriors.
arXiv Detail & Related papers (2020-11-08T17:09:37Z) - Mean-Field Approximation to Gaussian-Softmax Integral with Application
to Uncertainty Estimation [23.38076756988258]
We propose a new single-model based approach to quantify uncertainty in deep neural networks.
We use a mean-field approximation formula to compute an analytically intractable integral.
Empirically, the proposed approach performs competitively when compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-06-13T07:32:38Z) - Efficiently Sampling Functions from Gaussian Process Posteriors [76.94808614373609]
We propose an easy-to-use and general-purpose approach for fast posterior sampling.
We demonstrate how decoupled sample paths accurately represent Gaussian process posteriors at a fraction of the usual cost.
arXiv Detail & Related papers (2020-02-21T14:03:16Z) - Bayesian Deep Learning and a Probabilistic Perspective of Generalization [56.69671152009899]
We show that deep ensembles provide an effective mechanism for approximate Bayesian marginalization.
We also propose a related approach that further improves the predictive distribution by marginalizing within basins of attraction.
arXiv Detail & Related papers (2020-02-20T15:13:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.