Families of costs with zero and nonnegative MTW tensor in optimal
transport
- URL: http://arxiv.org/abs/2401.00953v1
- Date: Mon, 1 Jan 2024 20:33:27 GMT
- Title: Families of costs with zero and nonnegative MTW tensor in optimal
transport
- Authors: Du Nguyen
- Abstract summary: We compute explicitly the MTW tensor for the optimal transport problem on $mathbbRn$ with a cost function of form $mathsfc$.
We analyze the $sinh$-type hyperbolic cost, providing examples of $mathsfc$-type functions and divergence.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We compute explicitly the MTW tensor (or cross curvature) for the optimal
transport problem on $\mathbb{R}^n$ with a cost function of form $\mathsf{c}(x,
y) = \mathsf{u}(x^{\mathfrak{t}}y)$, where $\mathsf{u}$ is a scalar function
with inverse $\mathsf{s}$, $x^{\ft}y$ is a nondegenerate bilinear pairing of
vectors $x, y$ belonging to an open subset of $\mathbb{R}^n$. The condition
that the MTW-tensor vanishes on null vectors under the Kim-McCann metric is a
fourth-order nonlinear ODE, which could be reduced to a linear ODE of the form
$\mathsf{s}^{(2)} - S\mathsf{s}^{(1)} + P\mathsf{s} = 0$ with constant
coefficients $P$ and $S$. The resulting inverse functions include {\it Lambert}
and {\it generalized inverse hyperbolic\slash trigonometric} functions. The
square Euclidean metric and $\log$-type costs are equivalent to instances of
these solutions. The optimal map for the family is also explicit. For cost
functions of a similar form on a hyperboloid model of the hyperbolic space and
unit sphere, we also express this tensor in terms of algebraic expressions in
derivatives of $\mathsf{s}$ using the Gauss-Codazzi equation, obtaining new
families of strictly regular costs for these manifolds, including new families
of {\it power function costs}. We analyze the $\sinh$-type hyperbolic cost,
providing examples of $\mathsf{c}$-convex functions and divergence.
Related papers
- Quantum charges of harmonic oscillators [55.2480439325792]
We show that the energy eigenfunctions $psi_n$ with $nge 1$ are complex coordinates on orbifolds $mathbbR2/mathbbZ_n$.
We also discuss "antioscillators" with opposite quantum charges and the same positive energy.
arXiv Detail & Related papers (2024-04-02T09:16:18Z) - Provably learning a multi-head attention layer [55.2904547651831]
Multi-head attention layer is one of the key components of the transformer architecture that sets it apart from traditional feed-forward models.
In this work, we initiate the study of provably learning a multi-head attention layer from random examples.
We prove computational lower bounds showing that in the worst case, exponential dependence on $m$ is unavoidable.
arXiv Detail & Related papers (2024-02-06T15:39:09Z) - A Unified Framework for Uniform Signal Recovery in Nonlinear Generative
Compressed Sensing [68.80803866919123]
Under nonlinear measurements, most prior results are non-uniform, i.e., they hold with high probability for a fixed $mathbfx*$ rather than for all $mathbfx*$ simultaneously.
Our framework accommodates GCS with 1-bit/uniformly quantized observations and single index models as canonical examples.
We also develop a concentration inequality that produces tighter bounds for product processes whose index sets have low metric entropy.
arXiv Detail & Related papers (2023-09-25T17:54:19Z) - Noncompact uniform universal approximation [0.0]
The universal approximation theorem is generalised to uniform convergence on the (noncompact) input space $mathbbRn$.
All continuous functions that vanish at infinity can be uniformly approximated by neural networks.
arXiv Detail & Related papers (2023-08-07T08:54:21Z) - Learning Elastic Costs to Shape Monge Displacements [39.381326738705255]
Monge problem asks to find the most efficient way to map one distribution to the other.
elastic costs shape the textitdisplacements of Monge maps $T$.
We propose a numerical method to compute Monge maps that are provably optimal.
arXiv Detail & Related papers (2023-06-20T21:17:32Z) - Learning a Single Neuron with Adversarial Label Noise via Gradient
Descent [50.659479930171585]
We study a function of the form $mathbfxmapstosigma(mathbfwcdotmathbfx)$ for monotone activations.
The goal of the learner is to output a hypothesis vector $mathbfw$ that $F(mathbbw)=C, epsilon$ with high probability.
arXiv Detail & Related papers (2022-06-17T17:55:43Z) - Local approximation of operators [0.0]
We study the problem of determining the degree of approximation of a non-linear operator between metric spaces $mathfrakX$ and $mathfrakY$.
We establish constructive methods to do this efficiently, i.e., with the constants involved in the estimates on the approximation on $mathbbSd$ being $mathcalO(d1/6)$.
arXiv Detail & Related papers (2022-02-13T19:28:34Z) - Metric Hypertransformers are Universal Adapted Maps [4.83420384410068]
metric hypertransformers (MHTs) are capable of approxing any adapted map $F:mathscrXmathbbZrightarrow mathscrYmathbbZ$ with approximable complexity.
Our results provide the first (quantimat) universal approximation theorem compatible with any such $mathscrX$ and $mathscrY$.
arXiv Detail & Related papers (2022-01-31T10:03:46Z) - Random matrices in service of ML footprint: ternary random features with
no performance loss [55.30329197651178]
We show that the eigenspectrum of $bf K$ is independent of the distribution of the i.i.d. entries of $bf w$.
We propose a novel random technique, called Ternary Random Feature (TRF)
The computation of the proposed random features requires no multiplication and a factor of $b$ less bits for storage compared to classical random features.
arXiv Detail & Related papers (2021-10-05T09:33:49Z) - Near-Optimal SQ Lower Bounds for Agnostically Learning Halfspaces and
ReLUs under Gaussian Marginals [49.60752558064027]
We study the fundamental problems of agnostically learning halfspaces and ReLUs under Gaussian marginals.
Our lower bounds provide strong evidence that current upper bounds for these tasks are essentially best possible.
arXiv Detail & Related papers (2020-06-29T17:10:10Z) - A Canonical Transform for Strengthening the Local $L^p$-Type Universal
Approximation Property [4.18804572788063]
$Lp$-type universal approximation theorems guarantee that a given machine learning model class $mathscrFsubseteq C(mathbbRd,mathbbRD)$ is dense in $Lp_mu(mathbbRd,mathbbRD)$.
This paper proposes a generic solution to this approximation theoretic problem by introducing a canonical transformation which "upgrades $mathscrF$'s approximation property"
arXiv Detail & Related papers (2020-06-24T17:46:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.