Y-Diagonal Couplings: Approximating Posteriors with Conditional
Wasserstein Distances
- URL: http://arxiv.org/abs/2310.13433v1
- Date: Fri, 20 Oct 2023 11:46:05 GMT
- Title: Y-Diagonal Couplings: Approximating Posteriors with Conditional
Wasserstein Distances
- Authors: Jannis Chemseddine, Paul Hagemann, Christian Wald
- Abstract summary: In inverse problems, many conditional generative models approximate the posterior measure by minimizing a distance between the joint measure and its learned approximation.
We will introduce a conditional Wasserstein distance with a set of restricted couplings that equals the expected Wasserstein distance of the posteriors.
- Score: 0.4419843514606336
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In inverse problems, many conditional generative models approximate the
posterior measure by minimizing a distance between the joint measure and its
learned approximation. While this approach also controls the distance between
the posterior measures in the case of the Kullback Leibler divergence, it does
not hold true for the Wasserstein distance. We will introduce a conditional
Wasserstein distance with a set of restricted couplings that equals the
expected Wasserstein distance of the posteriors. By deriving its dual, we find
a rigorous way to motivate the loss of conditional Wasserstein GANs. We outline
conditions under which the vanilla and the conditional Wasserstein distance
coincide. Furthermore, we will show numerical examples where training with the
conditional Wasserstein distance yields favorable properties for posterior
sampling.
Related papers
- Conditional Wasserstein Distances with Applications in Bayesian OT Flow Matching [1.609940380983903]
In inverse problems, many conditional generative models approximate the posterior measure by minimizing a distance between the joint measure and its learned approximation.
We introduce a conditional Wasserstein distance via a set of restricted couplings that equals the expected Wasserstein distance of the posteriors.
We derive theoretical properties of the conditional Wasserstein distance, characterize the corresponding geodesics and velocity fields as well as the flow ODEs.
arXiv Detail & Related papers (2024-03-27T15:54:55Z) - Offline Imitation from Observation via Primal Wasserstein State Occupancy Matching [111.78179839856293]
We propose Primal Wasserstein DICE to minimize the primal Wasserstein distance between the learner and expert state occupancies.
Our framework is a generalization of SMODICE, and is the first work that unifies $f$-divergence and Wasserstein minimization.
arXiv Detail & Related papers (2023-11-02T15:41:57Z) - Federated Wasserstein Distance [16.892296712204597]
We introduce a principled way of computing the Wasserstein distance between two distributions in a federated manner.
We show how to estimate the Wasserstein distance between two samples stored and kept on different devices/clients whilst a central entity/server orchestrates the computations.
arXiv Detail & Related papers (2023-10-03T11:30:50Z) - Sliced Wasserstein Estimation with Control Variates [47.18652387199418]
Sliced Wasserstein (SW) distances between two probability measures are defined as the expectation of the Wasserstein distance between two one-dimensional projections.
Due to the intractability of the expectation, Monte Carlo integration is performed to estimate the value of the SW distance.
Despite having various variants, there has been no prior work that improves the Monte Carlo estimation scheme for the SW distance.
arXiv Detail & Related papers (2023-04-30T06:03:17Z) - Bounds on Wasserstein distances between continuous distributions using
independent samples [0.0]
We propose a linear combination of plug-in estimators for the squared 2-Wasserstein distance with a reduced bias that decays to zero with the true distance.
We apply it to approximately bound from above the 2-Wasserstein distance between the target and current distribution in Markov chain Monte Carlo.
arXiv Detail & Related papers (2022-03-22T11:26:18Z) - Wasserstein GANs Work Because They Fail (to Approximate the Wasserstein
Distance) [1.1470070927586016]
Wasserstein GANs are based on the idea of minimising the Wasserstein distance between a real and a generated distribution.
We provide an in-depth mathematical analysis of differences between the theoretical setup and the reality of training Wasserstein GANs.
arXiv Detail & Related papers (2021-03-02T12:30:25Z) - Rethinking Rotated Object Detection with Gaussian Wasserstein Distance
Loss [111.8807588392563]
Boundary discontinuity and its inconsistency to the final detection metric have been the bottleneck for rotating detection regression loss design.
We propose a novel regression loss based on Gaussian Wasserstein distance as a fundamental approach to solve the problem.
arXiv Detail & Related papers (2021-01-28T12:04:35Z) - Towards Generalized Implementation of Wasserstein Distance in GANs [46.79148259312607]
Wasserstein GANs (WGANs) built upon the Kantorovich-Rubinstein duality of Wasserstein distance.
In practice it does not always outperform other variants of GANs.
We propose a general WGAN training scheme named Sobolev Wasserstein GAN (SWGAN)
arXiv Detail & Related papers (2020-12-07T02:22:23Z) - Reweighting samples under covariate shift using a Wasserstein distance
criterion [0.0]
We study an optimal reweighting that minimizes the Wasserstein distance between the empirical measures of the two samples.
The consistency and some convergence rates in terms of expected Wasserstein distance are derived.
These results have some application in Uncertainty Quantification for decoupled estimation and in the bound of the generalization error for the Nearest Neighbor Regression.
arXiv Detail & Related papers (2020-10-19T07:23:55Z) - On Projection Robust Optimal Transport: Sample Complexity and Model
Misspecification [101.0377583883137]
Projection robust (PR) OT seeks to maximize the OT cost between two measures by choosing a $k$-dimensional subspace onto which they can be projected.
Our first contribution is to establish several fundamental statistical properties of PR Wasserstein distances.
Next, we propose the integral PR Wasserstein (IPRW) distance as an alternative to the PRW distance, by averaging rather than optimizing on subspaces.
arXiv Detail & Related papers (2020-06-22T14:35:33Z) - Disentangled Representation Learning with Wasserstein Total Correlation [90.44329632061076]
We introduce Wasserstein total correlation in both variational autoencoder and Wasserstein autoencoder settings to learn disentangled latent representations.
A critic is adversarially trained along with the main objective to estimate the Wasserstein total correlation term.
We show that the proposed approach has comparable performances on disentanglement with smaller sacrifices in reconstruction abilities.
arXiv Detail & Related papers (2019-12-30T05:31:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.