InfoOT: Information Maximizing Optimal Transport
- URL: http://arxiv.org/abs/2210.03164v2
- Date: Mon, 29 May 2023 18:28:29 GMT
- Title: InfoOT: Information Maximizing Optimal Transport
- Authors: Ching-Yao Chuang, Stefanie Jegelka, David Alvarez-Melis
- Abstract summary: InfoOT is an information-theoretic extension of optimal transport.
It maximizes the mutual information between domains while minimizing geometric distances.
This formulation yields a new projection method that is robust to outliers and generalizes to unseen samples.
- Score: 58.72713603244467
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Optimal transport aligns samples across distributions by minimizing the
transportation cost between them, e.g., the geometric distances. Yet, it
ignores coherence structure in the data such as clusters, does not handle
outliers well, and cannot integrate new data points. To address these
drawbacks, we propose InfoOT, an information-theoretic extension of optimal
transport that maximizes the mutual information between domains while
minimizing geometric distances. The resulting objective can still be formulated
as a (generalized) optimal transport problem, and can be efficiently solved by
projected gradient descent. This formulation yields a new projection method
that is robust to outliers and generalizes to unseen samples. Empirically,
InfoOT improves the quality of alignments across benchmarks in domain
adaptation, cross-domain retrieval, and single-cell alignment.
Related papers
- SPARE: Symmetrized Point-to-Plane Distance for Robust Non-Rigid Registration [76.40993825836222]
We propose SPARE, a novel formulation that utilizes a symmetrized point-to-plane distance for robust non-rigid registration.
The proposed method greatly improves the accuracy of non-rigid registration problems and maintains relatively high solution efficiency.
arXiv Detail & Related papers (2024-05-30T15:55:04Z) - Distributional Reduction: Unifying Dimensionality Reduction and Clustering with Gromov-Wasserstein [56.62376364594194]
Unsupervised learning aims to capture the underlying structure of potentially large and high-dimensional datasets.
In this work, we revisit these approaches under the lens of optimal transport and exhibit relationships with the Gromov-Wasserstein problem.
This unveils a new general framework, called distributional reduction, that recovers DR and clustering as special cases and allows addressing them jointly within a single optimization problem.
arXiv Detail & Related papers (2024-02-03T19:00:19Z) - Minimally Supervised Learning using Topological Projections in
Self-Organizing Maps [55.31182147885694]
We introduce a semi-supervised learning approach based on topological projections in self-organizing maps (SOMs)
Our proposed method first trains SOMs on unlabeled data and then a minimal number of available labeled data points are assigned to key best matching units (BMU)
Our results indicate that the proposed minimally supervised model significantly outperforms traditional regression techniques.
arXiv Detail & Related papers (2024-01-12T22:51:48Z) - Neural Optimal Transport with General Cost Functionals [66.41953045707172]
We introduce a novel neural network-based algorithm to compute optimal transport plans for general cost functionals.
As an application, we construct a cost functional to map data distributions while preserving the class-wise structure.
arXiv Detail & Related papers (2022-05-30T20:00:19Z) - Hierarchical Optimal Transport for Unsupervised Domain Adaptation [0.0]
We propose a novel approach for unsupervised domain adaptation, that relates notions of optimal transport, learning probability measures and unsupervised learning.
The proposed approach, HOT-DA, is based on a hierarchical formulation of optimal transport.
Experiments on a toy dataset with controllable complexity and two challenging visual adaptation datasets show the superiority of the proposed approach over the state-of-the-art.
arXiv Detail & Related papers (2021-12-03T18:37:23Z) - Do Neural Optimal Transport Solvers Work? A Continuous Wasserstein-2
Benchmark [133.46066694893318]
We evaluate the performance of neural network-based solvers for optimal transport.
We find that existing solvers do not recover optimal transport maps even though they perform well in downstream tasks.
arXiv Detail & Related papers (2021-06-03T15:59:28Z) - Unbalanced minibatch Optimal Transport; applications to Domain
Adaptation [8.889304968879163]
Optimal transport distances have found many applications in machine learning for their capacity to compare non-parametric probability distributions.
We argue that the same minibatch strategy coupled with unbalanced optimal transport can yield more robust behavior.
Our experimental study shows that in challenging problems associated to domain adaptation, the use of unbalanced optimal transport leads to significantly better results, competing with or surpassing recent baselines.
arXiv Detail & Related papers (2021-03-05T11:15:47Z) - Making transport more robust and interpretable by moving data through a
small number of anchor points [4.262693522975888]
We introduce a new form of structured OT that simultaneously learns low-dimensional structure in data while leveraging this structure to solve the alignment task.
Compared with OT, the resulting transport plan has better structural interpretability, highlighting the connections between individual data points and local geometry.
We show that our method can facilitate alignment in noisy settings and can be used to both correct and interpret domain shift.
arXiv Detail & Related papers (2020-12-21T18:59:47Z) - A contribution to Optimal Transport on incomparable spaces [4.873362301533825]
This thesis proposes to study the complex scenario in which the different data belong to incomparable spaces.
This thesis proposes a set of Optimal Transport tools for these different cases.
arXiv Detail & Related papers (2020-11-09T14:13:52Z) - An Information-Geometric Distance on the Space of Tasks [31.359578768463752]
This paper prescribes a distance between learning tasks modeled as joint distributions on data and labels.
We develop an algorithm to compute the distance which iteratively transports the marginal on the data of the source task to that of the target task.
We perform thorough empirical validation and analysis across diverse image classification datasets to show that the coupled transfer distance correlates strongly with the difficulty of fine-tuning.
arXiv Detail & Related papers (2020-11-01T19:48:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.