Semi-Discrete Optimal Transport: Hardness, Regularization and Numerical
Solution
- URL: http://arxiv.org/abs/2103.06263v1
- Date: Wed, 10 Mar 2021 18:53:59 GMT
- Title: Semi-Discrete Optimal Transport: Hardness, Regularization and Numerical
Solution
- Authors: Bahar Taskesen, Soroosh Shafieezadeh-Abadeh, Daniel Kuhn
- Abstract summary: We prove that computing the Wasserstein distance between a discrete probability measure supported on two points is already #P-hard.
We introduce a distributionally robust dual optimal transport problem whose objective function is smoothed with the most adverse disturbance distributions.
We show that smoothing the dual objective function is equivalent to regularizing the primal objective function.
- Score: 8.465228064780748
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semi-discrete optimal transport problems, which evaluate the Wasserstein
distance between a discrete and a generic (possibly non-discrete) probability
measure, are believed to be computationally hard. Even though such problems are
ubiquitous in statistics, machine learning and computer vision, however, this
perception has not yet received a theoretical justification. To fill this gap,
we prove that computing the Wasserstein distance between a discrete probability
measure supported on two points and the Lebesgue measure on the standard
hypercube is already #P-hard. This insight prompts us to seek approximate
solutions for semi-discrete optimal transport problems. We thus perturb the
underlying transportation cost with an additive disturbance governed by an
ambiguous probability distribution, and we introduce a distributionally robust
dual optimal transport problem whose objective function is smoothed with the
most adverse disturbance distributions from within a given ambiguity set. We
further show that smoothing the dual objective function is equivalent to
regularizing the primal objective function, and we identify several ambiguity
sets that give rise to several known and new regularization schemes. As a
byproduct, we discover an intimate relation between semi-discrete optimal
transport problems and discrete choice models traditionally studied in
psychology and economics. To solve the regularized optimal transport problems
efficiently, we use a stochastic gradient descent algorithm with imprecise
stochastic gradient oracles. A new convergence analysis reveals that this
algorithm improves the best known convergence guarantee for semi-discrete
optimal transport problems with entropic regularizers.
Related papers
- An Inexact Halpern Iteration with Application to Distributionally Robust
Optimization [9.529117276663431]
We investigate the inexact variants of the scheme in both deterministic and deterministic convergence settings.
We show that by choosing the inexactness appropriately, the inexact schemes admit an $O(k-1) convergence rate in terms of the (expected) residue norm.
arXiv Detail & Related papers (2024-02-08T20:12:47Z) - Moreau-Yoshida Variational Transport: A General Framework For Solving Regularized Distributional Optimization Problems [3.038642416291856]
We consider a general optimization problem of minimizing a composite objective functional defined over a class probability distributions.
We propose a novel method, dubbed as Moreau-Yoshida Variational Transport (MYVT), for solving the regularized distributional optimization problem.
arXiv Detail & Related papers (2023-07-31T01:14:42Z) - New Perspectives on Regularization and Computation in Optimal
Transport-Based Distributionally Robust Optimization [8.564319625930892]
We study optimal transport-based distributionally robust optimization problems where a fictitious adversary, often envisioned as nature, can choose the distribution of the uncertain problem parameters by a prescribed reference distribution at a finite transportation cost.
arXiv Detail & Related papers (2023-03-07T13:52:32Z) - A Short and General Duality Proof for Wasserstein Distributionally Robust Optimization [11.034091190797671]
We present a general duality result for Wasserstein distributionally robust optimization that holds for any Kantorovich transport cost, measurable loss function, and nominal probability distribution.
We demonstrate that the interchangeability principle holds if and only if certain measurable projection and weak measurable selection conditions are satisfied.
arXiv Detail & Related papers (2022-04-30T22:49:01Z) - Optimal variance-reduced stochastic approximation in Banach spaces [114.8734960258221]
We study the problem of estimating the fixed point of a contractive operator defined on a separable Banach space.
We establish non-asymptotic bounds for both the operator defect and the estimation error.
arXiv Detail & Related papers (2022-01-21T02:46:57Z) - Nearly Tight Convergence Bounds for Semi-discrete Entropic Optimal
Transport [0.483420384410068]
We derive nearly tight and non-asymptotic convergence bounds for solutions of entropic semi-discrete optimal transport.
Our results also entail a non-asymptotic and tight expansion of the difference between the entropic and the unregularized costs.
arXiv Detail & Related papers (2021-10-25T06:52:45Z) - Lifting the Convex Conjugate in Lagrangian Relaxations: A Tractable
Approach for Continuous Markov Random Fields [53.31927549039624]
We show that a piecewise discretization preserves better contrast from existing discretization problems.
We apply this theory to the problem of matching two images.
arXiv Detail & Related papers (2021-07-13T12:31:06Z) - Stochastic Gradient Descent-Ascent and Consensus Optimization for Smooth
Games: Convergence Analysis under Expected Co-coercivity [49.66890309455787]
We introduce the expected co-coercivity condition, explain its benefits, and provide the first last-iterate convergence guarantees of SGDA and SCO.
We prove linear convergence of both methods to a neighborhood of the solution when they use constant step-size.
Our convergence guarantees hold under the arbitrary sampling paradigm, and we give insights into the complexity of minibatching.
arXiv Detail & Related papers (2021-06-30T18:32:46Z) - High Probability Complexity Bounds for Non-Smooth Stochastic Optimization with Heavy-Tailed Noise [51.31435087414348]
It is essential to theoretically guarantee that algorithms provide small objective residual with high probability.
Existing methods for non-smooth convex optimization have complexity bounds with dependence on confidence level.
We propose novel stepsize rules for two methods with gradient clipping.
arXiv Detail & Related papers (2021-06-10T17:54:21Z) - Variational Transport: A Convergent Particle-BasedAlgorithm for Distributional Optimization [106.70006655990176]
A distributional optimization problem arises widely in machine learning and statistics.
We propose a novel particle-based algorithm, dubbed as variational transport, which approximately performs Wasserstein gradient descent.
We prove that when the objective function satisfies a functional version of the Polyak-Lojasiewicz (PL) (Polyak, 1963) and smoothness conditions, variational transport converges linearly.
arXiv Detail & Related papers (2020-12-21T18:33:13Z) - Continuous Regularized Wasserstein Barycenters [51.620781112674024]
We introduce a new dual formulation for the regularized Wasserstein barycenter problem.
We establish strong duality and use the corresponding primal-dual relationship to parametrize the barycenter implicitly using the dual potentials of regularized transport problems.
arXiv Detail & Related papers (2020-08-28T08:28:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.