Optimal transport map estimation in general function spaces
- URL: http://arxiv.org/abs/2212.03722v2
- Date: Tue, 2 Jan 2024 21:25:15 GMT
- Title: Optimal transport map estimation in general function spaces
- Authors: Vincent Divol, Jonathan Niles-Weed, Aram-Alexandre Pooladian
- Abstract summary: We study the problem of estimating a function $T$ given independent samples from a distribution $P$ and from the pushforward distribution $T_sharp P$.
This setting is motivated by applications in the sciences, where $T$ represents the evolution of a physical system over time.
We present a unified methodology for obtaining rates of estimation of optimal transport maps in general function spaces.
- Score: 17.323588442718926
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study the problem of estimating a function $T$ given independent samples
from a distribution $P$ and from the pushforward distribution $T_\sharp P$.
This setting is motivated by applications in the sciences, where $T$ represents
the evolution of a physical system over time, and in machine learning, where,
for example, $T$ may represent a transformation learned by a deep neural
network trained for a generative modeling task. To ensure identifiability, we
assume that $T = \nabla \varphi_0$ is the gradient of a convex function, in
which case $T$ is known as an \emph{optimal transport map}. Prior work has
studied the estimation of $T$ under the assumption that it lies in a H\"older
class, but general theory is lacking. We present a unified methodology for
obtaining rates of estimation of optimal transport maps in general function
spaces. Our assumptions are significantly weaker than those appearing in the
literature: we require only that the source measure $P$ satisfy a Poincar\'e
inequality and that the optimal map be the gradient of a smooth convex function
that lies in a space whose metric entropy can be controlled. As a special case,
we recover known estimation rates for H\"older transport maps, but also obtain
nearly sharp results in many settings not covered by prior work. For example,
we provide the first statistical rates of estimation when $P$ is the normal
distribution and the transport map is given by an infinite-width shallow neural
network.
Related papers
- Minimax Optimality of Score-based Diffusion Models: Beyond the Density Lower Bound Assumptions [11.222970035173372]
kernel-based score estimator achieves an optimal mean square error of $widetildeOleft(n-1 t-fracd+22(tfracd2 vee 1)right)
We show that a kernel-based score estimator achieves an optimal mean square error of $widetildeOleft(n-1/2 t-fracd4right)$ upper bound for the total variation error of the distribution of the sample generated by the diffusion model under a mere sub-Gaussian
arXiv Detail & Related papers (2024-02-23T20:51:31Z) - Normalizing flows as approximations of optimal transport maps via linear-control neural ODEs [49.1574468325115]
"Normalizing Flows" is related to the task of constructing invertible transport maps between probability measures by means of deep neural networks.
We consider the problem of recovering the $Wamma$-optimal transport map $T$ between absolutely continuous measures $mu,nuinmathcalP(mathbbRn)$ as the flow of a linear-control neural ODE.
arXiv Detail & Related papers (2023-11-02T17:17:03Z) - Minimax estimation of discontinuous optimal transport maps: The
semi-discrete case [14.333765302506658]
We consider the problem of estimating the optimal transport map between two probability distributions, $P$ and $Q$ in $mathbb Rd$.
We show that an estimator based on entropic optimal transport converges at the minimax-optimal rate $n-1/2$, independent of dimension.
arXiv Detail & Related papers (2023-01-26T18:41:38Z) - On the Identifiability and Estimation of Causal Location-Scale Noise
Models [122.65417012597754]
We study the class of location-scale or heteroscedastic noise models (LSNMs)
We show the causal direction is identifiable up to some pathological cases.
We propose two estimators for LSNMs: an estimator based on (non-linear) feature maps, and one based on neural networks.
arXiv Detail & Related papers (2022-10-13T17:18:59Z) - Supervised Training of Conditional Monge Maps [107.78770597815242]
Optimal transport (OT) theory describes general principles to define and select, among many possible choices, the most efficient way to map a probability measure onto another.
We introduce CondOT, a multi-task approach to estimate a family of OT maps conditioned on a context variable.
We demonstrate the ability of CondOT to infer the effect of an arbitrary combination of genetic or therapeutic perturbations on single cells.
arXiv Detail & Related papers (2022-06-28T19:34:44Z) - Approximate Function Evaluation via Multi-Armed Bandits [51.146684847667125]
We study the problem of estimating the value of a known smooth function $f$ at an unknown point $boldsymbolmu in mathbbRn$, where each component $mu_i$ can be sampled via a noisy oracle.
We design an instance-adaptive algorithm that learns to sample according to the importance of each coordinate, and with probability at least $1-delta$ returns an $epsilon$ accurate estimate of $f(boldsymbolmu)$.
arXiv Detail & Related papers (2022-03-18T18:50:52Z) - Locality defeats the curse of dimensionality in convolutional
teacher-student scenarios [69.2027612631023]
We show that locality is key in determining the learning curve exponent $beta$.
We conclude by proving, using a natural assumption, that performing kernel regression with a ridge that decreases with the size of the training set leads to similar learning curve exponents to those we obtain in the ridgeless case.
arXiv Detail & Related papers (2021-06-16T08:27:31Z) - Randomized Exploration for Reinforcement Learning with General Value
Function Approximation [122.70803181751135]
We propose a model-free reinforcement learning algorithm inspired by the popular randomized least squares value iteration (RLSVI) algorithm.
Our algorithm drives exploration by simply perturbing the training data with judiciously chosen i.i.d. scalar noises.
We complement the theory with an empirical evaluation across known difficult exploration tasks.
arXiv Detail & Related papers (2021-06-15T02:23:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.