$\alpha$-Geodesical Skew Divergence
- URL: http://arxiv.org/abs/2103.17060v2
- Date: Thu, 1 Apr 2021 05:40:47 GMT
- Title: $\alpha$-Geodesical Skew Divergence
- Authors: Masanari Kimura and Hideitsu Hino
- Abstract summary: The asymmetric skew divergence smooths one of the distributions by mixing it, to a degree determined by the parameter $lambda$, with the other distribution.
Such divergence is an approximation of the KL divergence that does not require the target distribution to be absolutely continuous with respect to the source distribution.
- Score: 5.3556221126231085
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The asymmetric skew divergence smooths one of the distributions by mixing it,
to a degree determined by the parameter $\lambda$, with the other distribution.
Such divergence is an approximation of the KL divergence that does not require
the target distribution to be absolutely continuous with respect to the source
distribution. In this paper, an information geometric generalization of the
skew divergence called the $\alpha$-geodesical skew divergence is proposed, and
its properties are studied.
Related papers
- Statistical and Geometrical properties of regularized Kernel Kullback-Leibler divergence [7.273481485032721]
We study the statistical and geometrical properties of the Kullback-Leibler divergence with kernel covariance operators introduced by Bach [2022]
Unlike the classical Kullback-Leibler (KL) divergence that involves density ratios, the KKL compares probability distributions through covariance operators (embeddings) in a reproducible kernel Hilbert space (RKHS)
This novel divergence hence shares parallel but different aspects with both the standard Kullback-Leibler between probability distributions and kernel embeddings metrics such as the maximum mean discrepancy.
arXiv Detail & Related papers (2024-08-29T14:01:30Z) - Adaptive Annealed Importance Sampling with Constant Rate Progress [68.8204255655161]
Annealed Importance Sampling (AIS) synthesizes weighted samples from an intractable distribution.
We propose the Constant Rate AIS algorithm and its efficient implementation for $alpha$-divergences.
arXiv Detail & Related papers (2023-06-27T08:15:28Z) - Wrapped Distributions on homogeneous Riemannian manifolds [58.720142291102135]
Control over distributions' properties, such as parameters, symmetry and modality yield a family of flexible distributions.
We empirically validate our approach by utilizing our proposed distributions within a variational autoencoder and a latent space network model.
arXiv Detail & Related papers (2022-04-20T21:25:21Z) - Non-Gaussian Component Analysis via Lattice Basis Reduction [56.98280399449707]
Non-Gaussian Component Analysis (NGCA) is a distribution learning problem.
We provide an efficient algorithm for NGCA in the regime that $A$ is discrete or nearly discrete.
arXiv Detail & Related papers (2021-12-16T18:38:02Z) - Spectral clustering under degree heterogeneity: a case for the random
walk Laplacian [83.79286663107845]
This paper shows that graph spectral embedding using the random walk Laplacian produces vector representations which are completely corrected for node degree.
In the special case of a degree-corrected block model, the embedding concentrates about K distinct points, representing communities.
arXiv Detail & Related papers (2021-05-03T16:36:27Z) - On the Robustness to Misspecification of $\alpha$-Posteriors and Their
Variational Approximations [12.52149409594807]
$alpha$-posteriors and their variational approximations distort standard posterior inference by downweighting the likelihood and introducing variational approximation errors.
We show that such distortions, if tuned appropriately, reduce the Kullback-Leibler (KL) divergence from the true, but perhaps infeasible, posterior distribution when there is potential parametric model misspecification.
arXiv Detail & Related papers (2021-04-16T19:11:53Z) - Annealed Importance Sampling with q-Paths [51.73925445218365]
Annealed importance sampling (AIS) is the gold standard for estimating partition functions or marginal likelihoods.
Existing literature has been primarily limited to the geometric mixture or moment-averaged paths associated with the exponential family and KL divergence.
We explore AIS using $q$-paths, which include the geometric path as a special case and are related to the homogeneous power mean, deformed exponential family, and $alpha$-divergence.
arXiv Detail & Related papers (2020-12-14T18:57:05Z) - Independent Gaussian Distributions Minimize the Kullback-Leibler (KL)
Divergence from Independent Gaussian Distributions [23.249999313567624]
This note is on a property of the Kullback-Leibler (KL) divergence.
The primary purpose of this note is for the referencing of papers that need to make use of this property entirely or partially.
arXiv Detail & Related papers (2020-11-04T22:05:45Z) - Linear Optimal Transport Embedding: Provable Wasserstein classification
for certain rigid transformations and perturbations [79.23797234241471]
Discriminating between distributions is an important problem in a number of scientific fields.
The Linear Optimal Transportation (LOT) embeds the space of distributions into an $L2$-space.
We demonstrate the benefits of LOT on a number of distribution classification problems.
arXiv Detail & Related papers (2020-08-20T19:09:33Z) - Kullback-Leibler divergence between quantum distributions, and its
upper-bound [1.2183405753834562]
This work presents an upper-bound to value that the Kullback-Leibler (KL) divergence can reach for a class of probability distributions called quantum distributions (QD)
The retrieving of an upper-bound for the entropic divergence is here shown to be possible under the condition that the compared distributions are quantum distributions over the same quantum value, thus they become comparable.
arXiv Detail & Related papers (2020-08-13T14:42:13Z) - Generalized Bayesian Cram\'{e}r-Rao Inequality via Information Geometry
of Relative $\alpha$-Entropy [17.746238062801293]
relative $alpha$-entropy is the R'enyi analog of relative entropy.
Recent information geometric investigations on this quantity have enabled the generalization of the Cram'er-Rao inequality.
We show that in the limiting case when the entropy order approaches unity, this framework reduces to the conventional Bayesian Cram'er-Rao inequality.
arXiv Detail & Related papers (2020-02-11T23:38:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.