Linear-Time Gromov Wasserstein Distances using Low Rank Couplings and
Costs
- URL: http://arxiv.org/abs/2106.01128v1
- Date: Wed, 2 Jun 2021 12:50:56 GMT
- Title: Linear-Time Gromov Wasserstein Distances using Low Rank Couplings and
Costs
- Authors: Meyer Scetbon, Gabriel Peyr\'e, Marco Cuturi
- Abstract summary: The ability to compare and align related datasets living in heterogeneous spaces plays an increasingly important role in machine learning.
The Gromov-Wasserstein (GW) formalism can help tackle this problem.
- Score: 45.87981728307819
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The ability to compare and align related datasets living in heterogeneous
spaces plays an increasingly important role in machine learning. The
Gromov-Wasserstein (GW) formalism can help tackle this problem. Its main goal
is to seek an assignment (more generally a coupling matrix) that can register
points across otherwise incomparable datasets. As a non-convex and quadratic
generalization of optimal transport (OT), GW is NP-hard. Yet, heuristics are
known to work reasonably well in practice, the state of the art approach being
to solve a sequence of nested regularized OT problems. While popular, that
heuristic remains too costly to scale, with cubic complexity in the number of
samples $n$. We show in this paper how a recent variant of the Sinkhorn
algorithm can substantially speed up the resolution of GW. That variant
restricts the set of admissible couplings to those admitting a low rank
factorization as the product of two sub-couplings. By updating alternatively
each sub-coupling, our algorithm computes a stationary point of the problem in
quadratic time with respect to the number of samples. When cost matrices have
themselves low rank, our algorithm has time complexity $\mathcal{O}(n)$. We
demonstrate the efficiency of our method on simulated and real data.
Related papers
- A Novel Normalized-Cut Solver with Nearest Neighbor Hierarchical
Initialization [107.07093621337084]
Normalized-Cut (N-Cut) is a famous model of spectral clustering.
Traditional N-Cut solvers are two-stage: 1) calculating the continuous spectral embedding of normalized Laplacian matrix; 2) discretization via $K$-means or spectral rotation.
We propose a novel N-Cut solver based on the famous coordinate descent method.
arXiv Detail & Related papers (2023-11-26T07:11:58Z) - Efficiently Learning One-Hidden-Layer ReLU Networks via Schur
Polynomials [50.90125395570797]
We study the problem of PAC learning a linear combination of $k$ ReLU activations under the standard Gaussian distribution on $mathbbRd$ with respect to the square loss.
Our main result is an efficient algorithm for this learning task with sample and computational complexity $(dk/epsilon)O(k)$, whereepsilon>0$ is the target accuracy.
arXiv Detail & Related papers (2023-07-24T14:37:22Z) - Efficient Strongly Polynomial Algorithms for Quantile Regression [12.772027028644041]
We revisit the classical technique of Quantile Regression (QR)
This paper proposes several strongly efficient classical algorithms for QR for various settings.
For two dimensional QR, making a connection to the geometric concept of $k$-set, we propose an algorithm with a worst-case time complexity of $mathcalO(n4/3 polylog(n))$.
We also propose a randomized divide-and-conquer algorithm -- RandomizedQR with an expected time complexity of $mathcalO(nlog2(n))$ for two dimensional
arXiv Detail & Related papers (2023-07-14T03:07:42Z) - Learning Transition Operators From Sparse Space-Time Samples [11.859913430860335]
We consider the nonlinear problem of learning a transition operator $mathbfA$ from partial observations at different times.
We show that no more than $mathcalOrn log(nT)$ space-time samples are sufficient to ensure accurate recovery of a rank-$r$ operator.
arXiv Detail & Related papers (2022-12-01T18:33:59Z) - Finite-Time Error Bounds for Greedy-GQ [20.51105692499517]
We show that Greedy-GQ algorithm converges fast as finite-time error.
Our analysis provides for faster convergence step-sizes for choosing step-sizes.
arXiv Detail & Related papers (2022-09-06T15:04:57Z) - Multi-block-Single-probe Variance Reduced Estimator for Coupled
Compositional Optimization [49.58290066287418]
We propose a novel method named Multi-block-probe Variance Reduced (MSVR) to alleviate the complexity of compositional problems.
Our results improve upon prior ones in several aspects, including the order of sample complexities and dependence on strongity.
arXiv Detail & Related papers (2022-07-18T12:03:26Z) - FANOK: Knockoffs in Linear Time [73.5154025911318]
We describe a series of algorithms that efficiently implement Gaussian model-X knockoffs to control the false discovery rate on large scale feature selection problems.
We test our methods on problems with $p$ as large as $500,000$.
arXiv Detail & Related papers (2020-06-15T21:55:34Z) - Fast and Robust Comparison of Probability Measures in Heterogeneous
Spaces [62.35667646858558]
We introduce the Anchor Energy (AE) and Anchor Wasserstein (AW) distances, which are respectively the energy and Wasserstein distances instantiated on such representations.
Our main contribution is to propose a sweep line algorithm to compute AE emphexactly in log-quadratic time, where a naive implementation would be cubic.
We show that AE and AW perform well in various experimental settings at a fraction of the computational cost of popular GW approximations.
arXiv Detail & Related papers (2020-02-05T03:09:23Z) - Learning Sparse Classifiers: Continuous and Mixed Integer Optimization
Perspectives [10.291482850329892]
Mixed integer programming (MIP) can be used to solve (to optimality) $ell_0$-regularized regression problems.
We propose two classes of scalable algorithms: an exact algorithm that can handlepapprox 50,000$ features in a few minutes, and approximate algorithms that can address instances with $papprox6$.
In addition, we present new estimation error bounds for $ell$-regularizeds.
arXiv Detail & Related papers (2020-01-17T18:47:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.