Wasserstein Distances Made Explainable: Insights into Dataset Shifts and Transport Phenomena
- URL: http://arxiv.org/abs/2505.06123v1
- Date: Fri, 09 May 2025 15:26:38 GMT
- Title: Wasserstein Distances Made Explainable: Insights into Dataset Shifts and Transport Phenomena
- Authors: Philip Naumann, Jacob Kauffmann, Grégoire Montavon,
- Abstract summary: Wasserstein distances provide a powerful framework for comparing data distributions.<n>We propose a novel solution based on Explainable AI that allows us to efficiently and accurately attribute Wasserstein distances to various data components.
- Score: 3.4991519098475843
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Wasserstein distances provide a powerful framework for comparing data distributions. They can be used to analyze processes over time or to detect inhomogeneities within data. However, simply calculating the Wasserstein distance or analyzing the corresponding transport map (or coupling) may not be sufficient for understanding what factors contribute to a high or low Wasserstein distance. In this work, we propose a novel solution based on Explainable AI that allows us to efficiently and accurately attribute Wasserstein distances to various data components, including data subgroups, input features, or interpretable subspaces. Our method achieves high accuracy across diverse datasets and Wasserstein distance specifications, and its practical utility is demonstrated in two use cases.
Related papers
- Wasserstein-based Kernels for Clustering: Application to Power Distribution Graphs [0.0]
This work explores kernel methods and Wasserstein distance metrics to develop a computationally tractable clustering framework.<n>The framework is flexible enough to be applied in various domains, such as graph analysis and image processing.<n>A case study involving two datasets of 879 and 34,920 power distribution graphs demonstrates the framework's effectiveness and efficiency.
arXiv Detail & Related papers (2025-03-18T15:40:55Z) - Fused Gromov-Wasserstein Variance Decomposition with Linear Optimal Transport [11.94799054956877]
We present a decomposition of Fr'echet variance of a set of measures in the 2-Wasserstein space, which allows one to compute the percentage of variance explained by LOT embeddings of those measures.
We also present several experiments that explore the relationship between the dimension of the LOT embedding, the percentage of variance explained, and the classification accuracy of machine learning classifiers built on the embedded data.
arXiv Detail & Related papers (2024-11-15T14:10:52Z) - Private Wasserstein Distance [6.015898117103069]
Wasserstein distance is a key metric for quantifying data divergence from a distributional perspective.<n>In this study, we explore the inherent triangular properties within the Wasserstein space, leading to a novel solution named TriangleWad.
arXiv Detail & Related papers (2024-04-10T06:58:58Z) - Linearized Wasserstein dimensionality reduction with approximation
guarantees [65.16758672591365]
LOT Wassmap is a computationally feasible algorithm to uncover low-dimensional structures in the Wasserstein space.
We show that LOT Wassmap attains correct embeddings and that the quality improves with increased sample size.
We also show how LOT Wassmap significantly reduces the computational cost when compared to algorithms that depend on pairwise distance computations.
arXiv Detail & Related papers (2023-02-14T22:12:16Z) - Wasserstein t-SNE [25.241296604908424]
We develop an approach for exploratory analysis of hierarchical datasets using the Wasserstein distance metric.
We use t-SNE to construct 2D embeddings of the units, based on the matrix of pairwise Wasserstein distances between them.
arXiv Detail & Related papers (2022-05-16T09:09:24Z) - Partial Wasserstein Covering [10.52782170493037]
We consider a general task called partial Wasserstein covering with the goal of emulating a large dataset.
We model this problem as a discrete optimization problem with partial Wasserstein divergence as an objective function.
We show that we can efficiently make two datasets similar in terms of partial Wasserstein divergence, including driving scene datasets.
arXiv Detail & Related papers (2021-06-02T01:48:41Z) - Ranking the information content of distance measures [61.754016309475745]
We introduce a statistical test that can assess the relative information retained when using two different distance measures.
This in turn allows finding the most informative distance measure out of a pool of candidates.
arXiv Detail & Related papers (2021-04-30T15:57:57Z) - Learning High Dimensional Wasserstein Geodesics [55.086626708837635]
We propose a new formulation and learning strategy for computing the Wasserstein geodesic between two probability distributions in high dimensions.
By applying the method of Lagrange multipliers to the dynamic formulation of the optimal transport (OT) problem, we derive a minimax problem whose saddle point is the Wasserstein geodesic.
We then parametrize the functions by deep neural networks and design a sample based bidirectional learning algorithm for training.
arXiv Detail & Related papers (2021-02-05T04:25:28Z) - Two-sample Test using Projected Wasserstein Distance [18.46110328123008]
We develop a projected Wasserstein distance for the two-sample test, a fundamental problem in statistics and machine learning.
A key contribution is to couple optimal projection to find the low dimensional linear mapping to maximize the Wasserstein distance between projected probability distributions.
arXiv Detail & Related papers (2020-10-22T18:08:58Z) - On Projection Robust Optimal Transport: Sample Complexity and Model
Misspecification [101.0377583883137]
Projection robust (PR) OT seeks to maximize the OT cost between two measures by choosing a $k$-dimensional subspace onto which they can be projected.
Our first contribution is to establish several fundamental statistical properties of PR Wasserstein distances.
Next, we propose the integral PR Wasserstein (IPRW) distance as an alternative to the PRW distance, by averaging rather than optimizing on subspaces.
arXiv Detail & Related papers (2020-06-22T14:35:33Z) - Augmented Sliced Wasserstein Distances [55.028065567756066]
We propose a new family of distance metrics, called augmented sliced Wasserstein distances (ASWDs)
ASWDs are constructed by first mapping samples to higher-dimensional hypersurfaces parameterized by neural networks.
Numerical results demonstrate that the ASWD significantly outperforms other Wasserstein variants for both synthetic and real-world problems.
arXiv Detail & Related papers (2020-06-15T23:00:08Z) - Projection Robust Wasserstein Distance and Riemannian Optimization [107.93250306339694]
We show that projection robustly solidstein (PRW) is a robust variant of Wasserstein projection (WPP)
This paper provides a first step into the computation of the PRW distance and provides the links between their theory and experiments on and real data.
arXiv Detail & Related papers (2020-06-12T20:40:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.