Bounds on Wasserstein distances between continuous distributions using
independent samples
- URL: http://arxiv.org/abs/2203.11627v1
- Date: Tue, 22 Mar 2022 11:26:18 GMT
- Title: Bounds on Wasserstein distances between continuous distributions using
independent samples
- Authors: Tam\'as Papp and Chris Sherlock
- Abstract summary: We propose a linear combination of plug-in estimators for the squared 2-Wasserstein distance with a reduced bias that decays to zero with the true distance.
We apply it to approximately bound from above the 2-Wasserstein distance between the target and current distribution in Markov chain Monte Carlo.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The plug-in estimator of the Wasserstein distance is known to be
conservative, however its usefulness is severely limited when the distributions
are similar as its bias does not decay to zero with the true Wasserstein
distance. We propose a linear combination of plug-in estimators for the squared
2-Wasserstein distance with a reduced bias that decays to zero with the true
distance. The new estimator is provably conservative provided one distribution
is appropriately overdispersed with respect the other, and is unbiased when the
distributions are equal. We apply it to approximately bound from above the
2-Wasserstein distance between the target and current distribution in Markov
chain Monte Carlo, running multiple identically distributed chains which start,
and remain, overdispersed with respect to the target. Our bound consistently
outperforms the current state-of-the-art bound, which uses coupling, improving
mixing time bounds by up to an order of magnitude.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - Y-Diagonal Couplings: Approximating Posteriors with Conditional
Wasserstein Distances [0.4419843514606336]
In inverse problems, many conditional generative models approximate the posterior measure by minimizing a distance between the joint measure and its learned approximation.
We will introduce a conditional Wasserstein distance with a set of restricted couplings that equals the expected Wasserstein distance of the posteriors.
arXiv Detail & Related papers (2023-10-20T11:46:05Z) - Fisher-Rao distance and pullback SPD cone distances between multivariate normal distributions [7.070726553564701]
We introduce a class of distances based on diffeomorphic embeddings of the normal manifold into a submanifold.
We show that the projective Hilbert distance on the cone yields a metric on the embedded normal submanifold.
We show how to use those distances in clustering tasks.
arXiv Detail & Related papers (2023-07-20T07:14:58Z) - Adaptive Annealed Importance Sampling with Constant Rate Progress [68.8204255655161]
Annealed Importance Sampling (AIS) synthesizes weighted samples from an intractable distribution.
We propose the Constant Rate AIS algorithm and its efficient implementation for $alpha$-divergences.
arXiv Detail & Related papers (2023-06-27T08:15:28Z) - Energy-Based Sliced Wasserstein Distance [47.18652387199418]
A key component of the sliced Wasserstein (SW) distance is the slicing distribution.
We propose to design the slicing distribution as an energy-based distribution that is parameter-free.
We then derive a novel sliced Wasserstein metric, energy-based sliced Waserstein (EBSW) distance.
arXiv Detail & Related papers (2023-04-26T14:28:45Z) - Hilbert Curve Projection Distance for Distribution Comparison [34.8765820950517]
We propose a novel metric, called Hilbert curve projection (HCP) distance, to measure the distance between two probability distributions.
We show that HCP distance is a proper metric and is well-defined for probability measures with bounded supports.
Experiments on both synthetic and real-world data show that our HCP distance works as an effective surrogate of the Wasserstein distance with low complexity.
arXiv Detail & Related papers (2022-05-30T12:40:32Z) - Density Ratio Estimation via Infinitesimal Classification [85.08255198145304]
We propose DRE-infty, a divide-and-conquer approach to reduce Density ratio estimation (DRE) to a series of easier subproblems.
Inspired by Monte Carlo methods, we smoothly interpolate between the two distributions via an infinite continuum of intermediate bridge distributions.
We show that our approach performs well on downstream tasks such as mutual information estimation and energy-based modeling on complex, high-dimensional datasets.
arXiv Detail & Related papers (2021-11-22T06:26:29Z) - Understanding Entropic Regularization in GANs [5.448283690603358]
We study the influence of regularization on the learned solution of Wasserstein distance.
We show that entropy regularization promotes the solution sparsification, while replacing the Wasserstein distance with the Sinkhorn divergence recovers the unregularized solution.
We conclude that these regularization techniques can improve the quality of the generator learned from empirical data for a large class of distributions.
arXiv Detail & Related papers (2021-11-02T06:08:16Z) - Continuous Regularized Wasserstein Barycenters [51.620781112674024]
We introduce a new dual formulation for the regularized Wasserstein barycenter problem.
We establish strong duality and use the corresponding primal-dual relationship to parametrize the barycenter implicitly using the dual potentials of regularized transport problems.
arXiv Detail & Related papers (2020-08-28T08:28:06Z) - Linear Optimal Transport Embedding: Provable Wasserstein classification
for certain rigid transformations and perturbations [79.23797234241471]
Discriminating between distributions is an important problem in a number of scientific fields.
The Linear Optimal Transportation (LOT) embeds the space of distributions into an $L2$-space.
We demonstrate the benefits of LOT on a number of distribution classification problems.
arXiv Detail & Related papers (2020-08-20T19:09:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.