Partial Optimal Transport with Applications on Positive-Unlabeled
Learning
- URL: http://arxiv.org/abs/2002.08276v2
- Date: Fri, 12 Jun 2020 09:48:54 GMT
- Title: Partial Optimal Transport with Applications on Positive-Unlabeled
Learning
- Authors: Laetitia Chapel, Mokhtar Z. Alaya, Gilles Gasso
- Abstract summary: We propose exact algorithms to solve Wasserstein and Gromov-Wasserstein problems.
This is the first application of optimal transport in this context.
- Score: 12.504473943407092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Classical optimal transport problem seeks a transportation map that preserves
the total mass betwenn two probability distributions, requiring their mass to
be the same. This may be too restrictive in certain applications such as color
or shape matching, since the distributions may have arbitrary masses and/or
that only a fraction of the total mass has to be transported. Several
algorithms have been devised for computing partial Wasserstein metrics that
rely on an entropic regularization, but when it comes with exact solutions,
almost no partial formulation of neither Wasserstein nor Gromov-Wasserstein are
available yet. This precludes from working with distributions that do not lie
in the same metric space or when invariance to rotation or translation is
needed. In this paper, we address the partial Wasserstein and
Gromov-Wasserstein problems and propose exact algorithms to solve them. We
showcase the new formulation in a positive-unlabeled (PU) learning application.
To the best of our knowledge, this is the first application of optimal
transport in this context and we first highlight that partial Wasserstein-based
metrics prove effective in usual PU learning settings. We then demonstrate that
partial Gromov-Wasserstein metrics is efficient in scenario where point clouds
come from different domains or have different features.
Related papers
- Fast Estimation of Wasserstein Distances via Regression on Sliced Wasserstein Distances [70.94157767200342]
We propose a fast estimation method based on regressing Wasserstein distance on sliced Wasserstein distances.<n>We show that accurate models can be learned from a small number of distribution pairs.<n>Our method consistently provides a better approximation of Wasserstein distance than the state-of-the-art Wasserstein embedding model, Wasserstein Wormhole.
arXiv Detail & Related papers (2025-09-24T19:30:53Z) - Repulsive Monte Carlo on the sphere for the sliced Wasserstein distance [4.052758394413725]
We consider the problem of computing the integral of a function on the unit sphere, in any dimension, using Monte Carlo methods.<n>Our guiding thread is the sliced Wasserstein distance between two measures on $mathbbRd$, which is precisely an integral on the $d$-dimensional sphere.
arXiv Detail & Related papers (2025-09-12T11:48:11Z) - Energy-Based Sliced Wasserstein Distance [47.18652387199418]
A key component of the sliced Wasserstein (SW) distance is the slicing distribution.
We propose to design the slicing distribution as an energy-based distribution that is parameter-free.
We then derive a novel sliced Wasserstein metric, energy-based sliced Waserstein (EBSW) distance.
arXiv Detail & Related papers (2023-04-26T14:28:45Z) - A new method for determining Wasserstein 1 optimal transport maps from
Kantorovich potentials, with deep learning applications [3.867363075280544]
Wasserstein 1 optimal transport maps provide a natural correspondence between points from two probability distributions.
We present an approach towards computing Wasserstein 1 optimal transport maps that relies only on Kantorovich potentials.
arXiv Detail & Related papers (2022-11-02T01:54:09Z) - Sliced Wasserstein Variational Inference [3.405431122165563]
We propose a new variational inference method by minimizing sliced Wasserstein distance, a valid metric arising from optimal transport.
Our approximation also does not require a tractable density function of variational distributions so that approximating families can be amortized by generators like neural networks.
arXiv Detail & Related papers (2022-07-26T20:51:51Z) - Near-optimal estimation of smooth transport maps with kernel
sums-of-squares [81.02564078640275]
Under smoothness conditions, the squared Wasserstein distance between two distributions could be efficiently computed with appealing statistical error upper bounds.
The object of interest for applications such as generative modeling is the underlying optimal transport map.
We propose the first tractable algorithm for which the statistical $L2$ error on the maps nearly matches the existing minimax lower-bounds for smooth map estimation.
arXiv Detail & Related papers (2021-12-03T13:45:36Z) - Non-asymptotic convergence bounds for Wasserstein approximation using
point clouds [0.0]
We show how to generate discrete data as if sampled from a model probability distribution.
We provide explicit upper bounds for the convergence-type algorithm.
arXiv Detail & Related papers (2021-06-15T06:53:08Z) - Continuous Wasserstein-2 Barycenter Estimation without Minimax
Optimization [94.18714844247766]
Wasserstein barycenters provide a geometric notion of the weighted average of probability measures based on optimal transport.
We present a scalable algorithm to compute Wasserstein-2 barycenters given sample access to the input measures.
arXiv Detail & Related papers (2021-02-02T21:01:13Z) - Variational Transport: A Convergent Particle-BasedAlgorithm for Distributional Optimization [106.70006655990176]
A distributional optimization problem arises widely in machine learning and statistics.
We propose a novel particle-based algorithm, dubbed as variational transport, which approximately performs Wasserstein gradient descent.
We prove that when the objective function satisfies a functional version of the Polyak-Lojasiewicz (PL) (Polyak, 1963) and smoothness conditions, variational transport converges linearly.
arXiv Detail & Related papers (2020-12-21T18:33:13Z) - Linear Optimal Transport Embedding: Provable Wasserstein classification
for certain rigid transformations and perturbations [79.23797234241471]
Discriminating between distributions is an important problem in a number of scientific fields.
The Linear Optimal Transportation (LOT) embeds the space of distributions into an $L2$-space.
We demonstrate the benefits of LOT on a number of distribution classification problems.
arXiv Detail & Related papers (2020-08-20T19:09:33Z) - Fast and Robust Comparison of Probability Measures in Heterogeneous
Spaces [62.35667646858558]
We introduce the Anchor Energy (AE) and Anchor Wasserstein (AW) distances, which are respectively the energy and Wasserstein distances instantiated on such representations.
Our main contribution is to propose a sweep line algorithm to compute AE emphexactly in log-quadratic time, where a naive implementation would be cubic.
We show that AE and AW perform well in various experimental settings at a fraction of the computational cost of popular GW approximations.
arXiv Detail & Related papers (2020-02-05T03:09:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.