Bregman Deviations of Generic Exponential Families
- URL: http://arxiv.org/abs/2201.07306v4
- Date: Thu, 13 Jul 2023 07:38:41 GMT
- Title: Bregman Deviations of Generic Exponential Families
- Authors: Sayak Ray Chowdhury, Patrick Saux, Odalric-Ambrym Maillard, Aditya
Gopalan
- Abstract summary: We revisit the method of mixture technique, also known as the Laplace method, to study the concentration phenomenon in generic exponential families.
Our bound is time-uniform and makes appear a quantity extending the classical information gain to exponential families.
- Score: 28.592975143984006
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We revisit the method of mixture technique, also known as the Laplace method,
to study the concentration phenomenon in generic exponential families.
Combining the properties of Bregman divergence associated with log-partition
function of the family with the method of mixtures for super-martingales, we
establish a generic bound controlling the Bregman divergence between the
parameter of the family and a finite sample estimate of the parameter. Our
bound is time-uniform and makes appear a quantity extending the classical
information gain to exponential families, which we call the Bregman information
gain. For the practitioner, we instantiate this novel bound to several
classical families, e.g., Gaussian, Bernoulli, Exponential, Weibull, Pareto,
Poisson and Chi-square yielding explicit forms of the confidence sets and the
Bregman information gain. We further numerically compare the resulting
confidence bounds to state-of-the-art alternatives for time-uniform
concentration and show that this novel method yields competitive results.
Finally, we highlight the benefit of our concentration bounds on some
illustrative applications.
Related papers
- Collaborative Heterogeneous Causal Inference Beyond Meta-analysis [68.4474531911361]
We propose a collaborative inverse propensity score estimator for causal inference with heterogeneous data.
Our method shows significant improvements over the methods based on meta-analysis when heterogeneity increases.
arXiv Detail & Related papers (2024-04-24T09:04:36Z) - Nonparametric Partial Disentanglement via Mechanism Sparsity: Sparse
Actions, Interventions and Sparse Temporal Dependencies [58.179981892921056]
This work introduces a novel principle for disentanglement we call mechanism sparsity regularization.
We propose a representation learning method that induces disentanglement by simultaneously learning the latent factors.
We show that the latent factors can be recovered by regularizing the learned causal graph to be sparse.
arXiv Detail & Related papers (2024-01-10T02:38:21Z) - Model-Based Uncertainty in Value Functions [89.31922008981735]
We focus on characterizing the variance over values induced by a distribution over MDPs.
Previous work upper bounds the posterior variance over values by solving a so-called uncertainty Bellman equation.
We propose a new uncertainty Bellman equation whose solution converges to the true posterior variance over values.
arXiv Detail & Related papers (2023-02-24T09:18:27Z) - Clustering above Exponential Families with Tempered Exponential Measures [28.532545355403123]
Link with exponential families has allowed $k$-means clustering to be generalized to a wide variety of data generating distributions.
Getting the framework to work above exponential families is important to lift roadblocks like the lack of robustness of some population minimizers carved in their axiomatization.
arXiv Detail & Related papers (2022-11-04T21:58:40Z) - Finite-Time Regret of Thompson Sampling Algorithms for Exponential
Family Multi-Armed Bandits [88.21288104408556]
We study the regret of Thompson sampling (TS) algorithms for exponential family bandits, where the reward distribution is from a one-dimensional exponential family.
We propose a Thompson sampling, termed Expulli, which uses a novel sampling distribution to avoid the under-estimation of the optimal arm.
arXiv Detail & Related papers (2022-06-07T18:08:21Z) - Robust Estimation for Nonparametric Families via Generative Adversarial
Networks [92.64483100338724]
We provide a framework for designing Generative Adversarial Networks (GANs) to solve high dimensional robust statistics problems.
Our work extend these to robust mean estimation, second moment estimation, and robust linear regression.
In terms of techniques, our proposed GAN losses can be viewed as a smoothed and generalized Kolmogorov-Smirnov distance.
arXiv Detail & Related papers (2022-02-02T20:11:33Z) - Kernel Deformed Exponential Families for Sparse Continuous Attention [76.61129971916702]
We show existence results for kernel exponential and deformed exponential families.
Experiments show that kernel deformed exponential families can attend to multiple compact regions of the data domain.
arXiv Detail & Related papers (2021-11-01T19:21:22Z) - Fast approximations of the Jeffreys divergence between univariate
Gaussian mixture models via exponential polynomial densities [16.069404547401373]
The Jeffreys divergence is a renown symmetrization of the statistical Kullback-Leibler which is often used in machine learning, signal processing, and information sciences.
We propose a simple yet fastarine to approximate the Jeffreys divergence between two GMMs of arbitrary number of components.
arXiv Detail & Related papers (2021-07-13T07:58:01Z) - Likelihood Ratio Exponential Families [43.98796887171374]
We use the geometric mixture path as an exponential family of distributions to analyze the thermodynamic variational objective (TVO)
We extend these likelihood ratio exponential families to include solutions to rate-distortion (RD) optimization, the information bottleneck (IB) method, and recent rate-distortion-classification approaches.
arXiv Detail & Related papers (2020-12-31T07:13:58Z) - Estimating means of bounded random variables by betting [39.98103047898974]
This paper derives confidence intervals (CI) and time-uniform confidence sequences (CS) for the classical problem of estimating an unknown mean from bounded observations.
We present a general approach for deriving concentration bounds, that can be seen as a generalization and improvement of the celebrated Chernoff method.
We show how to extend these ideas to sampling without replacement, another heavily studied problem.
arXiv Detail & Related papers (2020-10-19T17:22:03Z) - Optimal Bounds between $f$-Divergences and Integral Probability Metrics [8.401473551081748]
Families of $f$-divergences and Integral Probability Metrics are widely used to quantify similarity between probability distributions.
We systematically study the relationship between these two families from the perspective of convex duality.
We obtain new bounds while also recovering in a unified manner well-known results, such as Hoeffding's lemma.
arXiv Detail & Related papers (2020-06-10T17:39:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.