On the attainment of the Wasserstein--Cramer--Rao lower bound
- URL: http://arxiv.org/abs/2506.12732v2
- Date: Tue, 17 Jun 2025 15:04:48 GMT
- Title: On the attainment of the Wasserstein--Cramer--Rao lower bound
- Authors: Hayato Nishimori, Takeru Matsuda,
- Abstract summary: A Wasserstein analogue of the Cramer--Rao inequality has been developed using the Wasserstein information matrix (Otto metric)<n>This inequality provides a lower bound on the Wasserstein variance of an estimator, which quantifies its robustness against additive noise.<n>We investigate conditions for an estimator to attain the Wasserstein--Cramer--Rao lower bound, which we call the (asymptotic) Wasserstein efficiency.
- Score: 4.48890356952206
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, a Wasserstein analogue of the Cramer--Rao inequality has been developed using the Wasserstein information matrix (Otto metric). This inequality provides a lower bound on the Wasserstein variance of an estimator, which quantifies its robustness against additive noise. In this study, we investigate conditions for an estimator to attain the Wasserstein--Cramer--Rao lower bound (asymptotically), which we call the (asymptotic) Wasserstein efficiency. We show a condition under which Wasserstein efficient estimators exist for one-parameter statistical models. This condition corresponds to a recently proposed Wasserstein analogue of one-parameter exponential families (e-geodesics). We also show that the Wasserstein estimator, a Wasserstein analogue of the maximum likelihood estimator based on the Wasserstein score function, is asymptotically Wasserstein efficient in location-scale families.
Related papers
- Synthesis and Analysis of Data as Probability Measures with Entropy-Regularized Optimal Transport [12.573382512049053]
We consider synthesis and analysis of probability measures using the entropy-regularized Wasserstein-2 cost and its unbiased version, the Sinkhorn divergence.<n>We show that these coefficients, as well as the value of the barycenter functional, can be estimated from samples with dimension-independent rates of convergence.<n>We employ the barycentric coefficients as features for classification of corrupted point cloud data, and show that compared to neural network baselines, our approach is more efficient in small training data regimes.
arXiv Detail & Related papers (2025-01-13T16:16:53Z) - Variational Wasserstein gradient flow [9.901677207027806]
We propose a scalable proximal gradient type algorithm for Wasserstein gradient flow.
Our framework covers all the classical Wasserstein gradient flows including the heat equation and the porous medium equation.
arXiv Detail & Related papers (2021-12-04T20:27:31Z) - Non-asymptotic convergence bounds for Wasserstein approximation using
point clouds [0.0]
We show how to generate discrete data as if sampled from a model probability distribution.
We provide explicit upper bounds for the convergence-type algorithm.
arXiv Detail & Related papers (2021-06-15T06:53:08Z) - Wasserstein GANs Work Because They Fail (to Approximate the Wasserstein
Distance) [1.1470070927586016]
Wasserstein GANs are based on the idea of minimising the Wasserstein distance between a real and a generated distribution.
We provide an in-depth mathematical analysis of differences between the theoretical setup and the reality of training Wasserstein GANs.
arXiv Detail & Related papers (2021-03-02T12:30:25Z) - Learning High Dimensional Wasserstein Geodesics [55.086626708837635]
We propose a new formulation and learning strategy for computing the Wasserstein geodesic between two probability distributions in high dimensions.
By applying the method of Lagrange multipliers to the dynamic formulation of the optimal transport (OT) problem, we derive a minimax problem whose saddle point is the Wasserstein geodesic.
We then parametrize the functions by deep neural networks and design a sample based bidirectional learning algorithm for training.
arXiv Detail & Related papers (2021-02-05T04:25:28Z) - Continuous Wasserstein-2 Barycenter Estimation without Minimax
Optimization [94.18714844247766]
Wasserstein barycenters provide a geometric notion of the weighted average of probability measures based on optimal transport.
We present a scalable algorithm to compute Wasserstein-2 barycenters given sample access to the input measures.
arXiv Detail & Related papers (2021-02-02T21:01:13Z) - Two-sample Test using Projected Wasserstein Distance [18.46110328123008]
We develop a projected Wasserstein distance for the two-sample test, a fundamental problem in statistics and machine learning.
A key contribution is to couple optimal projection to find the low dimensional linear mapping to maximize the Wasserstein distance between projected probability distributions.
arXiv Detail & Related papers (2020-10-22T18:08:58Z) - Continuous Regularized Wasserstein Barycenters [51.620781112674024]
We introduce a new dual formulation for the regularized Wasserstein barycenter problem.
We establish strong duality and use the corresponding primal-dual relationship to parametrize the barycenter implicitly using the dual potentials of regularized transport problems.
arXiv Detail & Related papers (2020-08-28T08:28:06Z) - On Projection Robust Optimal Transport: Sample Complexity and Model
Misspecification [101.0377583883137]
Projection robust (PR) OT seeks to maximize the OT cost between two measures by choosing a $k$-dimensional subspace onto which they can be projected.
Our first contribution is to establish several fundamental statistical properties of PR Wasserstein distances.
Next, we propose the integral PR Wasserstein (IPRW) distance as an alternative to the PRW distance, by averaging rather than optimizing on subspaces.
arXiv Detail & Related papers (2020-06-22T14:35:33Z) - Projection Robust Wasserstein Distance and Riemannian Optimization [107.93250306339694]
We show that projection robustly solidstein (PRW) is a robust variant of Wasserstein projection (WPP)
This paper provides a first step into the computation of the PRW distance and provides the links between their theory and experiments on and real data.
arXiv Detail & Related papers (2020-06-12T20:40:22Z) - Disentangled Representation Learning with Wasserstein Total Correlation [90.44329632061076]
We introduce Wasserstein total correlation in both variational autoencoder and Wasserstein autoencoder settings to learn disentangled latent representations.
A critic is adversarially trained along with the main objective to estimate the Wasserstein total correlation term.
We show that the proposed approach has comparable performances on disentanglement with smaller sacrifices in reconstruction abilities.
arXiv Detail & Related papers (2019-12-30T05:31:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.