Differentiable Generalized Sliced Wasserstein Plans
- URL: http://arxiv.org/abs/2505.22049v1
- Date: Wed, 28 May 2025 07:18:08 GMT
- Title: Differentiable Generalized Sliced Wasserstein Plans
- Authors: Laetitia Chapel, Romain Tavenard, Samuel Vaiter,
- Abstract summary: Optimal Transport (OT) has attracted significant interest in the machine learning community.<n>A novel slicing scheme, dubbed min-SWGG, lifts a single one-dimensional plan back to the original multidimensional space.<n>We show that min-SWGG inherits typical limitations of slicing methods.<n>We propose a differentiable approximation scheme to efficiently identify the optimal slice, even in high-dimensional settings.
- Score: 10.764247782316984
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Optimal Transport (OT) has attracted significant interest in the machine learning community, not only for its ability to define meaningful distances between probability distributions -- such as the Wasserstein distance -- but also for its formulation of OT plans. Its computational complexity remains a bottleneck, though, and slicing techniques have been developed to scale OT to large datasets. Recently, a novel slicing scheme, dubbed min-SWGG, lifts a single one-dimensional plan back to the original multidimensional space, finally selecting the slice that yields the lowest Wasserstein distance as an approximation of the full OT plan. Despite its computational and theoretical advantages, min-SWGG inherits typical limitations of slicing methods: (i) the number of required slices grows exponentially with the data dimension, and (ii) it is constrained to linear projections. Here, we reformulate min-SWGG as a bilevel optimization problem and propose a differentiable approximation scheme to efficiently identify the optimal slice, even in high-dimensional settings. We furthermore define its generalized extension for accommodating to data living on manifolds. Finally, we demonstrate the practical value of our approach in various applications, including gradient flows on manifolds and high-dimensional spaces, as well as a novel sliced OT-based conditional flow matching for image generation -- where fast computation of transport plans is essential.
Related papers
- Geometric Operator Learning with Optimal Transport [77.16909146519227]
We propose integrating optimal transport (OT) into operator learning for partial differential equations (PDEs) on complex geometries.<n>For 3D simulations focused on surfaces, our OT-based neural operator embeds the surface geometry into a 2D parameterized latent space.<n> Experiments with Reynolds-averaged Navier-Stokes equations (RANS) on the ShapeNet-Car and DrivAerNet-Car datasets show that our method achieves better accuracy and also reduces computational expenses.
arXiv Detail & Related papers (2025-07-26T21:28:25Z) - Constrained Sliced Wasserstein Embedding [15.569545184712942]
We introduce a constrained learning approach to optimize the slicing directions for SW distances.<n>We demonstrate how this constrained slicing approach can be applied to pool high-dimensional embeddings into fixed-length permutation-invariant representations.
arXiv Detail & Related papers (2025-06-02T19:43:40Z) - Hierarchical Refinement: Optimal Transport to Infinity and Beyond [1.8749305679160366]
Optimal transport (OT) has enjoyed great success in machine-learning as a principled way to align datasets via a least-cost correspondence.<n>Sinkhorn has quadratic space complexity in the number of points, limiting the scalability to larger datasets.<n>We derive an algorithm that dynamically constructs a multiscale partition of a dataset using low-rank OT subproblems.
arXiv Detail & Related papers (2025-03-04T22:00:12Z) - Expected Sliced Transport Plans [9.33181953215826]
We propose a "lifting" operation to extend one-dimensional optimal transport plans back to the original space of the measures.
We prove that using the EST plan to weight the sum of the individual Euclidean costs for moving from one point to another results in a valid metric between the input discrete probability measures.
arXiv Detail & Related papers (2024-10-16T02:44:36Z) - Fast Optimal Transport through Sliced Wasserstein Generalized Geodesics [14.259614797710224]
min-SWGG is a proxy of the squared WD based on an optimal one-dimensional projection of the two input distributions.
We show that min-SWGG is an upper bound of WD and that it has a complexity similar to Sliced-Wasserstein.
Empirical evidences support the benefits of min-SWGG in various contexts.
arXiv Detail & Related papers (2023-07-04T15:20:41Z) - Linearized Wasserstein dimensionality reduction with approximation
guarantees [65.16758672591365]
LOT Wassmap is a computationally feasible algorithm to uncover low-dimensional structures in the Wasserstein space.
We show that LOT Wassmap attains correct embeddings and that the quality improves with increased sample size.
We also show how LOT Wassmap significantly reduces the computational cost when compared to algorithms that depend on pairwise distance computations.
arXiv Detail & Related papers (2023-02-14T22:12:16Z) - InfoOT: Information Maximizing Optimal Transport [58.72713603244467]
InfoOT is an information-theoretic extension of optimal transport.
It maximizes the mutual information between domains while minimizing geometric distances.
This formulation yields a new projection method that is robust to outliers and generalizes to unseen samples.
arXiv Detail & Related papers (2022-10-06T18:55:41Z) - Near-optimal estimation of smooth transport maps with kernel
sums-of-squares [81.02564078640275]
Under smoothness conditions, the squared Wasserstein distance between two distributions could be efficiently computed with appealing statistical error upper bounds.
The object of interest for applications such as generative modeling is the underlying optimal transport map.
We propose the first tractable algorithm for which the statistical $L2$ error on the maps nearly matches the existing minimax lower-bounds for smooth map estimation.
arXiv Detail & Related papers (2021-12-03T13:45:36Z) - Efficient estimates of optimal transport via low-dimensional embeddings [0.0]
Optimal transport distances (OT) have been widely used in recent work in Machine Learning as ways to compare probability distributions.
Recent work by Paty et al., 2019, aims specifically at reducing this cost by computing OT using low-rank projections of the data.
We extend this approach and show that one can approximate OT distances by using more general families of maps provided they are 1-Lipschitz.
arXiv Detail & Related papers (2021-11-08T21:22:51Z) - Generative Modeling with Optimal Transport Maps [83.59805931374197]
Optimal Transport (OT) has become a powerful tool for large-scale generative modeling tasks.
We show that the OT map itself can be used as a generative model, providing comparable performance.
arXiv Detail & Related papers (2021-10-06T18:17:02Z) - On Projection Robust Optimal Transport: Sample Complexity and Model
Misspecification [101.0377583883137]
Projection robust (PR) OT seeks to maximize the OT cost between two measures by choosing a $k$-dimensional subspace onto which they can be projected.
Our first contribution is to establish several fundamental statistical properties of PR Wasserstein distances.
Next, we propose the integral PR Wasserstein (IPRW) distance as an alternative to the PRW distance, by averaging rather than optimizing on subspaces.
arXiv Detail & Related papers (2020-06-22T14:35:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.