Sliced Wasserstein Estimation with Control Variates
- URL: http://arxiv.org/abs/2305.00402v2
- Date: Mon, 19 Feb 2024 00:06:49 GMT
- Title: Sliced Wasserstein Estimation with Control Variates
- Authors: Khai Nguyen and Nhat Ho
- Abstract summary: Sliced Wasserstein (SW) distances between two probability measures are defined as the expectation of the Wasserstein distance between two one-dimensional projections.
Due to the intractability of the expectation, Monte Carlo integration is performed to estimate the value of the SW distance.
Despite having various variants, there has been no prior work that improves the Monte Carlo estimation scheme for the SW distance.
- Score: 47.18652387199418
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The sliced Wasserstein (SW) distances between two probability measures are
defined as the expectation of the Wasserstein distance between two
one-dimensional projections of the two measures. The randomness comes from a
projecting direction that is used to project the two input measures to one
dimension. Due to the intractability of the expectation, Monte Carlo
integration is performed to estimate the value of the SW distance. Despite
having various variants, there has been no prior work that improves the Monte
Carlo estimation scheme for the SW distance in terms of controlling its
variance. To bridge the literature on variance reduction and the literature on
the SW distance, we propose computationally efficient control variates to
reduce the variance of the empirical estimation of the SW distance. The key
idea is to first find Gaussian approximations of projected one-dimensional
measures, then we utilize the closed-form of the Wasserstein-2 distance between
two Gaussian distributions to design the control variates. In particular, we
propose using a lower bound and an upper bound of the Wasserstein-2 distance
between two fitted Gaussians as two computationally efficient control variates.
We empirically show that the proposed control variate estimators can help to
reduce the variance considerably when comparing measures over images and
point-clouds. Finally, we demonstrate the favorable performance of the proposed
control variate estimators in gradient flows to interpolate between two
point-clouds and in deep generative modeling on standard image datasets, such
as CIFAR10 and CelebA.
Related papers
- Relative-Translation Invariant Wasserstein Distance [82.6068808353647]
We introduce a new family of distances, relative-translation invariant Wasserstein distances ($RW_p$)
We show that $RW_p distances are also real distance metrics defined on the quotient set $mathcalP_p(mathbbRn)/sim$ invariant to distribution translations.
arXiv Detail & Related papers (2024-09-04T03:41:44Z) - Sliced-Wasserstein Estimation with Spherical Harmonics as Control Variates [17.237390976128097]
Sliced-Wasserstein (SW) distance between probability measures is defined as the average of the Wasserstein resulting for associated one-dimensional projections.
Spherical harmonics are distances on the sphere that form an orthonormal basis of the set of square-integrable functions on the sphere.
An improved rate of convergence, compared to Monte Carlo, is established for general measures.
arXiv Detail & Related papers (2024-02-02T15:22:06Z) - Sliced Wasserstein with Random-Path Projecting Directions [49.802024788196434]
We propose an optimization-free slicing distribution that provides a fast sampling for the Monte Carlo estimation of expectation.
We derive the random-path slicing distribution (RPSD) and two variants of sliced Wasserstein, i.e., the Random-Path Projection Sliced Wasserstein (RPSW) and the Importance Weighted Random-Path Projection Sliced Wasserstein (IWRPSW)
arXiv Detail & Related papers (2024-01-29T04:59:30Z) - Mutual Wasserstein Discrepancy Minimization for Sequential
Recommendation [82.0801585843835]
We propose a novel self-supervised learning framework based on Mutual WasserStein discrepancy minimization MStein for the sequential recommendation.
We also propose a novel contrastive learning loss based on Wasserstein Discrepancy Measurement.
arXiv Detail & Related papers (2023-01-28T13:38:48Z) - Markovian Sliced Wasserstein Distances: Beyond Independent Projections [51.80527230603978]
We introduce a new family of SW distances, named Markovian sliced Wasserstein (MSW) distance, which imposes a first-order Markov structure on projecting directions.
We compare distances with previous SW variants in various applications such as flows, color transfer, and deep generative modeling to demonstrate the favorable performance of MSW.
arXiv Detail & Related papers (2023-01-10T01:58:15Z) - Fast Approximation of the Sliced-Wasserstein Distance Using
Concentration of Random Projections [19.987683989865708]
The Sliced-Wasserstein distance (SW) is being increasingly used in machine learning applications.
We propose a new perspective to approximate SW by making use of the concentration of measure phenomenon.
Our method does not require sampling a number of random projections, and is therefore both accurate and easy to use compared to the usual Monte Carlo approximation.
arXiv Detail & Related papers (2021-06-29T13:56:19Z) - Distributional Sliced Embedding Discrepancy for Incomparable
Distributions [22.615156512223766]
Gromov-Wasserstein (GW) distance is a key tool for manifold learning and cross-domain learning.
We propose a novel approach for comparing two computation distributions, that hinges on the idea of distributional slicing, embeddings, and on computing the closed-form Wasserstein distance between the sliced distributions.
arXiv Detail & Related papers (2021-06-04T15:11:30Z) - Two-sample Test using Projected Wasserstein Distance [18.46110328123008]
We develop a projected Wasserstein distance for the two-sample test, a fundamental problem in statistics and machine learning.
A key contribution is to couple optimal projection to find the low dimensional linear mapping to maximize the Wasserstein distance between projected probability distributions.
arXiv Detail & Related papers (2020-10-22T18:08:58Z) - On Projection Robust Optimal Transport: Sample Complexity and Model
Misspecification [101.0377583883137]
Projection robust (PR) OT seeks to maximize the OT cost between two measures by choosing a $k$-dimensional subspace onto which they can be projected.
Our first contribution is to establish several fundamental statistical properties of PR Wasserstein distances.
Next, we propose the integral PR Wasserstein (IPRW) distance as an alternative to the PRW distance, by averaging rather than optimizing on subspaces.
arXiv Detail & Related papers (2020-06-22T14:35:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.