Sliced Wasserstein Variational Inference
- URL: http://arxiv.org/abs/2207.13177v1
- Date: Tue, 26 Jul 2022 20:51:51 GMT
- Title: Sliced Wasserstein Variational Inference
- Authors: Mingxuan Yi and Song Liu
- Abstract summary: We propose a new variational inference method by minimizing sliced Wasserstein distance, a valid metric arising from optimal transport.
Our approximation also does not require a tractable density function of variational distributions so that approximating families can be amortized by generators like neural networks.
- Score: 3.405431122165563
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Variational Inference approximates an unnormalized distribution via the
minimization of Kullback-Leibler (KL) divergence. Although this divergence is
efficient for computation and has been widely used in applications, it suffers
from some unreasonable properties. For example, it is not a proper metric,
i.e., it is non-symmetric and does not preserve the triangle inequality. On the
other hand, optimal transport distances recently have shown some advantages
over KL divergence. With the help of these advantages, we propose a new
variational inference method by minimizing sliced Wasserstein distance, a valid
metric arising from optimal transport. This sliced Wasserstein distance can be
approximated simply by running MCMC but without solving any optimization
problem. Our approximation also does not require a tractable density function
of variational distributions so that approximating families can be amortized by
generators like neural networks. Furthermore, we provide an analysis of the
theoretical properties of our method. Experiments on synthetic and real data
are illustrated to show the performance of the proposed method.
Related papers
- Energy-Based Sliced Wasserstein Distance [47.18652387199418]
A key component of the sliced Wasserstein (SW) distance is the slicing distribution.
We propose to design the slicing distribution as an energy-based distribution that is parameter-free.
We then derive a novel sliced Wasserstein metric, energy-based sliced Waserstein (EBSW) distance.
arXiv Detail & Related papers (2023-04-26T14:28:45Z) - Kernel-based off-policy estimation without overlap: Instance optimality
beyond semiparametric efficiency [53.90687548731265]
We study optimal procedures for estimating a linear functional based on observational data.
For any convex and symmetric function class $mathcalF$, we derive a non-asymptotic local minimax bound on the mean-squared error.
arXiv Detail & Related papers (2023-01-16T02:57:37Z) - Statistical and Topological Properties of Gaussian Smoothed Sliced
Probability Divergences [9.08047281767226]
We show that smoothing and slicing preserve the metric property and the weak topology.
We also provide results on the sample complexity of such divergences.
arXiv Detail & Related papers (2021-10-20T12:21:32Z) - On the Robustness to Misspecification of $\alpha$-Posteriors and Their
Variational Approximations [12.52149409594807]
$alpha$-posteriors and their variational approximations distort standard posterior inference by downweighting the likelihood and introducing variational approximation errors.
We show that such distortions, if tuned appropriately, reduce the Kullback-Leibler (KL) divergence from the true, but perhaps infeasible, posterior distribution when there is potential parametric model misspecification.
arXiv Detail & Related papers (2021-04-16T19:11:53Z) - Continuous Wasserstein-2 Barycenter Estimation without Minimax
Optimization [94.18714844247766]
Wasserstein barycenters provide a geometric notion of the weighted average of probability measures based on optimal transport.
We present a scalable algorithm to compute Wasserstein-2 barycenters given sample access to the input measures.
arXiv Detail & Related papers (2021-02-02T21:01:13Z) - Variational Transport: A Convergent Particle-BasedAlgorithm for Distributional Optimization [106.70006655990176]
A distributional optimization problem arises widely in machine learning and statistics.
We propose a novel particle-based algorithm, dubbed as variational transport, which approximately performs Wasserstein gradient descent.
We prove that when the objective function satisfies a functional version of the Polyak-Lojasiewicz (PL) (Polyak, 1963) and smoothness conditions, variational transport converges linearly.
arXiv Detail & Related papers (2020-12-21T18:33:13Z) - The Unbalanced Gromov Wasserstein Distance: Conic Formulation and
Relaxation [0.0]
Comparing metric measure spaces (i.e. a metric space endowed with aprobability distribution) is at the heart of many machine learning problems.
The most popular distance between such metric spaces is the metric measureGro-Wasserstein (GW) distance of which is a quadratic.
The GW formulation alleviates the comparison of metric spaces equipped with arbitrary positive measures up to isometries.
arXiv Detail & Related papers (2020-09-09T12:38:14Z) - On Projection Robust Optimal Transport: Sample Complexity and Model
Misspecification [101.0377583883137]
Projection robust (PR) OT seeks to maximize the OT cost between two measures by choosing a $k$-dimensional subspace onto which they can be projected.
Our first contribution is to establish several fundamental statistical properties of PR Wasserstein distances.
Next, we propose the integral PR Wasserstein (IPRW) distance as an alternative to the PRW distance, by averaging rather than optimizing on subspaces.
arXiv Detail & Related papers (2020-06-22T14:35:33Z) - Projection Robust Wasserstein Distance and Riemannian Optimization [107.93250306339694]
We show that projection robustly solidstein (PRW) is a robust variant of Wasserstein projection (WPP)
This paper provides a first step into the computation of the PRW distance and provides the links between their theory and experiments on and real data.
arXiv Detail & Related papers (2020-06-12T20:40:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.