Schoenberg-Rao distances: Entropy-based and geometry-aware statistical
Hilbert distances
- URL: http://arxiv.org/abs/2002.08345v2
- Date: Tue, 28 Apr 2020 13:19:50 GMT
- Title: Schoenberg-Rao distances: Entropy-based and geometry-aware statistical
Hilbert distances
- Authors: Ga\"etan Hadjeres and Frank Nielsen
- Abstract summary: We study a class of statistical Hilbert distances that we term the Schoenberg-Rao distances.
We derive novel closed-form distances between mixtures of Gaussian distributions.
Our method constitutes a practical alternative to Wasserstein distances and we illustrate its efficiency on a broad range of machine learning tasks.
- Score: 12.729120803225065
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Distances between probability distributions that take into account the
geometry of their sample space,like the Wasserstein or the Maximum Mean
Discrepancy (MMD) distances have received a lot of attention in machine
learning as they can, for instance, be used to compare probability
distributions with disjoint supports. In this paper, we study a class of
statistical Hilbert distances that we term the Schoenberg-Rao distances, a
generalization of the MMD that allows one to consider a broader class of
kernels, namely the conditionally negative semi-definite kernels. In
particular, we introduce a principled way to construct such kernels and derive
novel closed-form distances between mixtures of Gaussian distributions. These
distances, derived from the concave Rao's quadratic entropy, enjoy nice
theoretical properties and possess interpretable hyperparameters which can be
tuned for specific applications. Our method constitutes a practical alternative
to Wasserstein distances and we illustrate its efficiency on a broad range of
machine learning tasks such as density estimation, generative modeling and
mixture simplification.
Related papers
- Exact mean and variance of the squared Hellinger distance for random density matrices [11.495104812547021]
The Hellinger distance between quantum states is a significant measure in quantum information theory.
We propose an approximation for the corresponding probability density function based on the gamma distribution.
arXiv Detail & Related papers (2024-09-22T18:56:49Z) - Weighted Riesz Particles [0.0]
We consider the target distribution as a mapping where the infinite-dimensional space of the parameters consists of a number of deterministic submanifolds.
We study the properties of the point, called Riesz, and embed it into sequential MCMC.
We find that there will be higher acceptance rates with fewer evaluations.
arXiv Detail & Related papers (2023-12-01T14:36:46Z) - Scaling Riemannian Diffusion Models [68.52820280448991]
We show that our method enables us to scale to high dimensional tasks on nontrivial manifold.
We model QCD densities on $SU(n)$ lattices and contrastively learned embeddings on high dimensional hyperspheres.
arXiv Detail & Related papers (2023-10-30T21:27:53Z) - Targeted Separation and Convergence with Kernel Discrepancies [61.973643031360254]
kernel-based discrepancy measures are required to (i) separate a target P from other probability measures or (ii) control weak convergence to P.
In this article we derive new sufficient and necessary conditions to ensure (i) and (ii)
For MMDs on separable metric spaces, we characterize those kernels that separate Bochner embeddable measures and introduce simple conditions for separating all measures with unbounded kernels.
arXiv Detail & Related papers (2022-09-26T16:41:16Z) - Cycle Consistent Probability Divergences Across Different Spaces [38.43511529063335]
Discrepancy measures between probability distributions are at the core of statistical inference and machine learning.
This work proposes a novel unbalanced Monge optimal transport formulation for matching, up to isometries, distributions on different spaces.
arXiv Detail & Related papers (2021-11-22T16:35:58Z) - Density Ratio Estimation via Infinitesimal Classification [85.08255198145304]
We propose DRE-infty, a divide-and-conquer approach to reduce Density ratio estimation (DRE) to a series of easier subproblems.
Inspired by Monte Carlo methods, we smoothly interpolate between the two distributions via an infinite continuum of intermediate bridge distributions.
We show that our approach performs well on downstream tasks such as mutual information estimation and energy-based modeling on complex, high-dimensional datasets.
arXiv Detail & Related papers (2021-11-22T06:26:29Z) - A Note on Optimizing Distributions using Kernel Mean Embeddings [94.96262888797257]
Kernel mean embeddings represent probability measures by their infinite-dimensional mean embeddings in a reproducing kernel Hilbert space.
We show that when the kernel is characteristic, distributions with a kernel sum-of-squares density are dense.
We provide algorithms to optimize such distributions in the finite-sample setting.
arXiv Detail & Related papers (2021-06-18T08:33:45Z) - Optimal oracle inequalities for solving projected fixed-point equations [53.31620399640334]
We study methods that use a collection of random observations to compute approximate solutions by searching over a known low-dimensional subspace of the Hilbert space.
We show how our results precisely characterize the error of a class of temporal difference learning methods for the policy evaluation problem with linear function approximation.
arXiv Detail & Related papers (2020-12-09T20:19:32Z) - Probing symmetries of quantum many-body systems through gap ratio
statistics [0.0]
We extend the study of the gap ratio distribution P(r) to the case where discrete symmetries are present.
We present a large set of applications in many-body physics, ranging from quantum clock models and anyonic chains to periodically-driven spin systems.
arXiv Detail & Related papers (2020-08-25T17:11:40Z) - Generalized Sliced Distances for Probability Distributions [47.543990188697734]
We introduce a broad family of probability metrics, coined as Generalized Sliced Probability Metrics (GSPMs)
GSPMs are rooted in the generalized Radon transform and come with a unique geometric interpretation.
We consider GSPM-based gradient flows for generative modeling applications and show that under mild assumptions, the gradient flow converges to the global optimum.
arXiv Detail & Related papers (2020-02-28T04:18:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.