A Family of Pairwise Multi-Marginal Optimal Transports that Define a
Generalized Metric
- URL: http://arxiv.org/abs/2001.11114v6
- Date: Thu, 22 Dec 2022 19:27:53 GMT
- Title: A Family of Pairwise Multi-Marginal Optimal Transports that Define a
Generalized Metric
- Authors: Liang Mi, Azadeh Sheikholeslami, and Jos\'e Bento
- Abstract summary: Multi-marginal OT (MMOT) generalizes OT to simultaneously transporting multiple distributions.
We prove new generalized metric properties for a family of pairwise MMOTs.
We illustrate the superiority of our MMOTs over other generalized metrics, and over non-metrics in both synthetic and real tasks.
- Score: 2.650860836597657
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Optimal transport (OT) problem is rapidly finding its way into machine
learning. Favoring its use are its metric properties. Many problems admit
solutions with guarantees only for objects embedded in metric spaces, and the
use of non-metrics can complicate solving them. Multi-marginal OT (MMOT)
generalizes OT to simultaneously transporting multiple distributions. It
captures important relations that are missed if the transport only involves two
distributions. Research on MMOT, however, has been focused on its existence,
uniqueness, practical algorithms, and the choice of cost functions. There is a
lack of discussion on the metric properties of MMOT, which limits its
theoretical and practical use. Here, we prove new generalized metric properties
for a family of pairwise MMOTs. We first explain the difficulty of proving this
via two negative results. Afterward, we prove the MMOTs' metric properties.
Finally, we show that the generalized triangle inequality of this family of
MMOTs cannot be improved. We illustrate the superiority of our MMOTs over other
generalized metrics, and over non-metrics in both synthetic and real tasks.
Related papers
- Mixture of Parrots: Experts improve memorization more than reasoning [72.445819694797]
We show that as we increase the number of experts, the memorization performance consistently increases while the reasoning capabilities saturate.
We find that increasing the number of experts helps solve knowledge-intensive tasks, but fails to yield the same benefits for reasoning tasks.
arXiv Detail & Related papers (2024-10-24T17:54:41Z) - Tempered Calculus for ML: Application to Hyperbolic Model Embedding [70.61101116794549]
Most mathematical distortions used in ML are fundamentally integral in nature.
In this paper, we unveil a grounded theory and tools which can help improve these distortions to better cope with ML requirements.
We show how to apply it to a problem that has recently gained traction in ML: hyperbolic embeddings with a "cheap" and accurate encoding along the hyperbolic vsean scale.
arXiv Detail & Related papers (2024-02-06T17:21:06Z) - Enriching Disentanglement: From Logical Definitions to Quantitative Metrics [59.12308034729482]
Disentangling the explanatory factors in complex data is a promising approach for data-efficient representation learning.
We establish relationships between logical definitions and quantitative metrics to derive theoretically grounded disentanglement metrics.
We empirically demonstrate the effectiveness of the proposed metrics by isolating different aspects of disentangled representations.
arXiv Detail & Related papers (2023-05-19T08:22:23Z) - Joint Metrics Matter: A Better Standard for Trajectory Forecasting [67.1375677218281]
Multi-modal trajectory forecasting methods evaluate using single-agent metrics (marginal metrics)
Only focusing on marginal metrics can lead to unnatural predictions, such as colliding trajectories or diverging trajectories for people who are clearly walking together as a group.
We present the first comprehensive evaluation of state-of-the-art trajectory forecasting methods with respect to multi-agent metrics (joint metrics): JADE, JFDE, and collision rate.
arXiv Detail & Related papers (2023-05-10T16:27:55Z) - Correcting Flaws in Common Disentanglement Metrics [44.937838134027714]
In this paper, we identify two failings of existing metrics, which mean they can assign a high score to a model which is still entangled.
We then consider the task of compositional generalization.
Unlike prior works, we treat this as a classification problem, which allows us to use it to measure the disentanglement ability of the encoder.
We show that performance on this task is (a) generally quite poor, (b) correlated with most disentanglement metrics, and (c) most strongly correlated with our newly proposed metrics.
arXiv Detail & Related papers (2023-04-05T09:43:58Z) - Adaptive Log-Euclidean Metrics for SPD Matrix Learning [73.12655932115881]
We propose Adaptive Log-Euclidean Metrics (ALEMs), which extend the widely used Log-Euclidean Metric (LEM)
The experimental and theoretical results demonstrate the merit of the proposed metrics in improving the performance of SPD neural networks.
arXiv Detail & Related papers (2023-03-26T18:31:52Z) - Goodness of Fit Metrics for Multi-class Predictor [0.0]
Several metrics are commonly used to measure fit goodness.
A leading constraint at least in emphreal world multi-class problems is imbalanced data.
We suggest generalizing Matthew's correlation coefficient into multi-dimensions.
arXiv Detail & Related papers (2022-08-11T06:07:29Z) - MMD-Regularized Unbalanced Optimal Transport [0.0]
We study the unbalanced optimal transport (UOT) problem, where the marginal constraints are enforced using Maximum Mean Discrepancy (MMD) regularization.
Our work is motivated by the observation that the literature on UOT is focused on regularization based on $phi$-divergence.
Despite the popularity of MMD, its role as a regularizer in the context of UOT seems less understood.
arXiv Detail & Related papers (2020-11-10T09:32:50Z) - Project and Forget: Solving Large-Scale Metric Constrained Problems [7.381113319198104]
Given a set of dissimilarity measurements amongst data points, determining what metric representation is most "consistent" with the input measurements is a key step in many machine learning algorithms.
Existing methods are restricted to specific kinds of metrics or small problem sizes because of the large number of metric constraints in such problems.
In this paper, we provide an active set, Project and Forget, that uses Bregman projections, to solve metric constrained problems with many (possibly exponentially) inequality constraints.
arXiv Detail & Related papers (2020-05-08T04:50:54Z) - Generalized Sliced Distances for Probability Distributions [47.543990188697734]
We introduce a broad family of probability metrics, coined as Generalized Sliced Probability Metrics (GSPMs)
GSPMs are rooted in the generalized Radon transform and come with a unique geometric interpretation.
We consider GSPM-based gradient flows for generative modeling applications and show that under mild assumptions, the gradient flow converges to the global optimum.
arXiv Detail & Related papers (2020-02-28T04:18:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.