Kantorovich Strikes Back! Wasserstein GANs are not Optimal Transport?
- URL: http://arxiv.org/abs/2206.07767v1
- Date: Wed, 15 Jun 2022 19:07:46 GMT
- Title: Kantorovich Strikes Back! Wasserstein GANs are not Optimal Transport?
- Authors: Alexander Korotin, Alexander Kolesov, Evgeny Burnaev
- Abstract summary: Wasserstein Generative Adversarial Networks (WGANs) are the popular generative models built on the theory of Optimal Transport (OT) and the Kantorovich duality.
Despite the success of WGANs, it is still unclear how well the underlying OT dual solvers approximate the OT cost (Wasserstein-1 distance, $mathbbW_1$) and the OT gradient needed to update the generator.
We construct 1-Lipschitz functions and use them to build ray monotone transport plans. This strategy yields pairs of continuous benchmark distributions with the analytically known OT plan, OT cost and OT
- Score: 138.1080446991979
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Wasserstein Generative Adversarial Networks (WGANs) are the popular
generative models built on the theory of Optimal Transport (OT) and the
Kantorovich duality. Despite the success of WGANs, it is still unclear how well
the underlying OT dual solvers approximate the OT cost (Wasserstein-1 distance,
$\mathbb{W}_{1}$) and the OT gradient needed to update the generator. In this
paper, we address these questions. We construct 1-Lipschitz functions and use
them to build ray monotone transport plans. This strategy yields pairs of
continuous benchmark distributions with the analytically known OT plan, OT cost
and OT gradient in high-dimensional spaces such as spaces of images. We
thoroughly evaluate popular WGAN dual form solvers (gradient penalty, spectral
normalization, entropic regularization, etc.) using these benchmark pairs. Even
though these solvers perform well in WGANs, none of them faithfully compute
$\mathbb{W}_{1}$ in high dimensions. Nevertheless, many provide a meaningful
approximation of the OT gradient. These observations suggest that these solvers
should not be treated as good estimators of $\mathbb{W}_{1}$, but to some
extent they indeed can be used in variational problems requiring the
minimization of $\mathbb{W}_{1}$.
Related papers
- Fast and scalable Wasserstein-1 neural optimal transport solver for single-cell perturbation prediction [55.89763969583124]
Optimal transport theory provides a principled framework for constructing such mappings.
We propose a novel optimal transport solver based on Wasserstein-1.
Our experiments demonstrate that the proposed solver can mimic the $W$ OT solvers in finding a unique and monotonic" map on 2D datasets.
arXiv Detail & Related papers (2024-11-01T14:23:19Z) - Semi-Discrete Optimal Transport: Nearly Minimax Estimation With Stochastic Gradient Descent and Adaptive Entropic Regularization [38.67914746910537]
We prove an $mathcalO(t-1)$ lower bound rate for the OT map, using the similarity between Laguerre cells estimation and density support estimation.
To nearly achieve the desired fast rate, we design an entropic regularization scheme decreasing with the number of samples.
arXiv Detail & Related papers (2024-05-23T11:46:03Z) - A Specialized Semismooth Newton Method for Kernel-Based Optimal
Transport [92.96250725599958]
Kernel-based optimal transport (OT) estimators offer an alternative, functional estimation procedure to address OT problems from samples.
We show that our SSN method achieves a global convergence rate of $O (1/sqrtk)$, and a local quadratic convergence rate under standard regularity conditions.
arXiv Detail & Related papers (2023-10-21T18:48:45Z) - Unbalanced Low-rank Optimal Transport Solvers [38.79369155558385]
We propose algorithms to implement extensions for the linear OT problem and its Fused-Gromov-Wasserstein generalization.
The goal of this paper is to merge these two strains, to achieve the promise of textitboth versatile/scalable unbalanced/low-rank OT solvers.
arXiv Detail & Related papers (2023-05-31T10:39:51Z) - Sparsity-Constrained Optimal Transport [27.76137474217754]
Regularized optimal transport is now increasingly used as a loss or as a matching layer in neural networks.
We propose a new approach for OT with explicit cardinality constraints on the transportation plan.
Our method can be thought as a middle ground between unregularized OT (recovered in the case $k$) and quadratically-regularized OT (recovered when $k$ is large enough)
arXiv Detail & Related papers (2022-09-30T13:39:47Z) - Provably Efficient Offline Reinforcement Learning with Trajectory-Wise
Reward [66.81579829897392]
We propose a novel offline reinforcement learning algorithm called Pessimistic vAlue iteRaTion with rEward Decomposition (PARTED)
PARTED decomposes the trajectory return into per-step proxy rewards via least-squares-based reward redistribution, and then performs pessimistic value based on the learned proxy reward.
To the best of our knowledge, PARTED is the first offline RL algorithm that is provably efficient in general MDP with trajectory-wise reward.
arXiv Detail & Related papers (2022-06-13T19:11:22Z) - On Unbalanced Optimal Transport: Gradient Methods, Sparsity and
Approximation Error [18.19398247972205]
We study the Unbalanced Optimal Transport (UOT) between two measures of possibly different masses with at most $n$ components.
We propose a novel algorithm based on Gradient Extrapolation Method (GEM-UOT) to find an $varepsilon$-approximate solution to the UOT problem.
arXiv Detail & Related papers (2022-02-08T03:22:39Z) - Generative Modeling with Optimal Transport Maps [83.59805931374197]
Optimal Transport (OT) has become a powerful tool for large-scale generative modeling tasks.
We show that the OT map itself can be used as a generative model, providing comparable performance.
arXiv Detail & Related papers (2021-10-06T18:17:02Z) - GO Hessian for Expectation-Based Objectives [73.06986780804269]
GO gradient was proposed recently for expectation-based objectives $mathbbE_q_boldsymbolboldsymbolgamma(boldsymboly) [f(boldsymboly)]$.
Based on the GO gradient, we present for $mathbbE_q_boldsymbolboldsymbolgamma(boldsymboly) [f(boldsymboly)]$ an unbiased low-variance Hessian estimator, named GO Hessian.
arXiv Detail & Related papers (2020-06-16T02:20:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.